[ 533.344360] env[61985]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61985) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 533.344713] env[61985]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61985) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 533.344750] env[61985]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61985) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 533.345092] env[61985]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 533.440065] env[61985]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61985) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 533.450031] env[61985]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61985) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 534.050123] env[61985]: INFO nova.virt.driver [None req-22254cbd-d3ed-4f69-aad2-0bbf158f516b None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 534.121480] env[61985]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.121643] env[61985]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.121764] env[61985]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61985) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 537.268301] env[61985]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-927ac807-ea81-4b06-a194-4c0b070efde9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.283774] env[61985]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61985) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 537.283898] env[61985]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-28543c8b-ba30-4c5c-b5e3-36af387d4015 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.311303] env[61985]: INFO oslo_vmware.api [-] Successfully established new session; session ID is b4b2e. [ 537.311437] env[61985]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.190s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.311944] env[61985]: INFO nova.virt.vmwareapi.driver [None req-22254cbd-d3ed-4f69-aad2-0bbf158f516b None None] VMware vCenter version: 7.0.3 [ 537.315281] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da700287-9043-48d0-b42b-e112162c3a0f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.336499] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7834e9-742b-4c63-900d-43ac84d36973 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.342453] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67971322-a599-4a10-b86c-184bebd3f849 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.349011] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4357296-0d95-4161-9024-b2d1c0943c4c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.361894] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fe4548-60ad-47d4-be52-fefc6651eab4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.367607] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6aca58-219b-47e0-8074-d39ae08c4149 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.397267] env[61985]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-24f59d11-a452-4a08-832a-228e391b45d8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.402168] env[61985]: DEBUG nova.virt.vmwareapi.driver [None req-22254cbd-d3ed-4f69-aad2-0bbf158f516b None None] Extension org.openstack.compute already exists. {{(pid=61985) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 537.404823] env[61985]: INFO nova.compute.provider_config [None req-22254cbd-d3ed-4f69-aad2-0bbf158f516b None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 537.908065] env[61985]: DEBUG nova.context [None req-22254cbd-d3ed-4f69-aad2-0bbf158f516b None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),9eb1e3bd-81f4-4880-8c45-81c3bb246663(cell1) {{(pid=61985) load_cells /opt/stack/nova/nova/context.py:464}} [ 537.910241] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.910479] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.911172] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.911613] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Acquiring lock "9eb1e3bd-81f4-4880-8c45-81c3bb246663" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.911807] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Lock "9eb1e3bd-81f4-4880-8c45-81c3bb246663" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.912840] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Lock "9eb1e3bd-81f4-4880-8c45-81c3bb246663" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.933235] env[61985]: INFO dbcounter [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Registered counter for database nova_cell0 [ 537.941488] env[61985]: INFO dbcounter [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Registered counter for database nova_cell1 [ 537.944935] env[61985]: DEBUG oslo_db.sqlalchemy.engines [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61985) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 537.945291] env[61985]: DEBUG oslo_db.sqlalchemy.engines [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61985) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 537.950212] env[61985]: ERROR nova.db.main.api [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.950212] env[61985]: result = function(*args, **kwargs) [ 537.950212] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.950212] env[61985]: return func(*args, **kwargs) [ 537.950212] env[61985]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 537.950212] env[61985]: result = fn(*args, **kwargs) [ 537.950212] env[61985]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 537.950212] env[61985]: return f(*args, **kwargs) [ 537.950212] env[61985]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 537.950212] env[61985]: return db.service_get_minimum_version(context, binaries) [ 537.950212] env[61985]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 537.950212] env[61985]: _check_db_access() [ 537.950212] env[61985]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 537.950212] env[61985]: stacktrace = ''.join(traceback.format_stack()) [ 537.950212] env[61985]: [ 537.950925] env[61985]: ERROR nova.db.main.api [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.950925] env[61985]: result = function(*args, **kwargs) [ 537.950925] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.950925] env[61985]: return func(*args, **kwargs) [ 537.950925] env[61985]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 537.950925] env[61985]: result = fn(*args, **kwargs) [ 537.950925] env[61985]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 537.950925] env[61985]: return f(*args, **kwargs) [ 537.950925] env[61985]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 537.950925] env[61985]: return db.service_get_minimum_version(context, binaries) [ 537.950925] env[61985]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 537.950925] env[61985]: _check_db_access() [ 537.950925] env[61985]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 537.950925] env[61985]: stacktrace = ''.join(traceback.format_stack()) [ 537.950925] env[61985]: [ 537.951409] env[61985]: WARNING nova.objects.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 537.951587] env[61985]: WARNING nova.objects.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Failed to get minimum service version for cell 9eb1e3bd-81f4-4880-8c45-81c3bb246663 [ 537.951993] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Acquiring lock "singleton_lock" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.952178] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Acquired lock "singleton_lock" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.952424] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Releasing lock "singleton_lock" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.952752] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Full set of CONF: {{(pid=61985) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 537.952896] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ******************************************************************************** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 537.953037] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Configuration options gathered from: {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 537.953177] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 537.953368] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 537.953501] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ================================================================================ {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 537.953719] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] allow_resize_to_same_host = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.953894] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] arq_binding_timeout = 300 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.954037] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] backdoor_port = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.954173] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] backdoor_socket = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.954342] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] block_device_allocate_retries = 60 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.954597] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] block_device_allocate_retries_interval = 3 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.954788] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cert = self.pem {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.954987] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.955149] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute_monitors = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.955326] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] config_dir = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.955499] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] config_drive_format = iso9660 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.955641] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.955852] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] config_source = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.956032] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] console_host = devstack {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.956226] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] control_exchange = nova {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.956392] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cpu_allocation_ratio = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.956560] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] daemon = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.956733] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] debug = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.956895] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] default_access_ip_network_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.957077] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] default_availability_zone = nova {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.957242] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] default_ephemeral_format = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.957422] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] default_green_pool_size = 1000 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.957677] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.957846] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] default_schedule_zone = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.958015] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] disk_allocation_ratio = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.958185] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] enable_new_services = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.958368] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] enabled_apis = ['osapi_compute'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.958560] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] enabled_ssl_apis = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.958724] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] flat_injected = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.958884] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] force_config_drive = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.959055] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] force_raw_images = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.959230] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] graceful_shutdown_timeout = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.959391] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] heal_instance_info_cache_interval = 60 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.959609] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] host = cpu-1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.959845] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.960071] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] initial_disk_allocation_ratio = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.960290] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] initial_ram_allocation_ratio = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.960469] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.960639] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] instance_build_timeout = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.960803] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] instance_delete_interval = 300 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.960976] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] instance_format = [instance: %(uuid)s] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.961161] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] instance_name_template = instance-%08x {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.961325] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] instance_usage_audit = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.961498] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] instance_usage_audit_period = month {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.961669] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.961842] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] instances_path = /opt/stack/data/nova/instances {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.962014] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] internal_service_availability_zone = internal {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.962180] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] key = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.962341] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] live_migration_retry_count = 30 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.962513] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] log_color = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.962681] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] log_config_append = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.962851] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.963023] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] log_dir = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.963189] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] log_file = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.963323] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] log_options = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.963484] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] log_rotate_interval = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.963656] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] log_rotate_interval_type = days {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.963826] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] log_rotation_type = none {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.963962] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.964107] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.964279] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.964448] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.964598] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.964760] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] long_rpc_timeout = 1800 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.964922] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] max_concurrent_builds = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.965096] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] max_concurrent_live_migrations = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.965260] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] max_concurrent_snapshots = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.965420] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] max_local_block_devices = 3 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.965581] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] max_logfile_count = 30 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.965742] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] max_logfile_size_mb = 200 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.965901] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] maximum_instance_delete_attempts = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.966154] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] metadata_listen = 0.0.0.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.966343] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] metadata_listen_port = 8775 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.966519] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] metadata_workers = 2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.966693] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] migrate_max_retries = -1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.966859] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] mkisofs_cmd = genisoimage {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.967082] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] my_block_storage_ip = 10.180.1.21 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.967221] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] my_ip = 10.180.1.21 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.967420] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] network_allocate_retries = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.967603] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.967781] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] osapi_compute_listen = 0.0.0.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.967955] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] osapi_compute_listen_port = 8774 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.968178] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] osapi_compute_unique_server_name_scope = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.968360] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] osapi_compute_workers = 2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.968527] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] password_length = 12 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.968692] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] periodic_enable = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.968859] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] periodic_fuzzy_delay = 60 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.969040] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] pointer_model = usbtablet {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.969215] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] preallocate_images = none {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.969377] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] publish_errors = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.969509] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] pybasedir = /opt/stack/nova {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.969670] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ram_allocation_ratio = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.969860] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] rate_limit_burst = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.970046] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] rate_limit_except_level = CRITICAL {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.970225] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] rate_limit_interval = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.970418] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] reboot_timeout = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.970532] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] reclaim_instance_interval = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.970689] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] record = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.970859] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] reimage_timeout_per_gb = 60 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.971038] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] report_interval = 120 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.971202] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] rescue_timeout = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.971365] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] reserved_host_cpus = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.971527] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] reserved_host_disk_mb = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.971690] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] reserved_host_memory_mb = 512 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.971849] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] reserved_huge_pages = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.972024] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] resize_confirm_window = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.972183] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] resize_fs_using_block_device = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.972346] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] resume_guests_state_on_host_boot = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.972516] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.972679] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] rpc_response_timeout = 60 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.972842] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] run_external_periodic_tasks = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.973019] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] running_deleted_instance_action = reap {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.973186] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] running_deleted_instance_poll_interval = 1800 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.973350] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] running_deleted_instance_timeout = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.973512] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler_instance_sync_interval = 120 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.973681] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_down_time = 720 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.973854] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] servicegroup_driver = db {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.974025] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] shell_completion = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.974190] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] shelved_offload_time = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.974355] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] shelved_poll_interval = 3600 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.974519] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] shutdown_timeout = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.974681] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] source_is_ipv6 = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.974838] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ssl_only = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.975093] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.975265] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] sync_power_state_interval = 600 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.975428] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] sync_power_state_pool_size = 1000 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.975598] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] syslog_log_facility = LOG_USER {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.975758] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] tempdir = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.975915] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] timeout_nbd = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.976100] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] transport_url = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.976353] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] update_resources_interval = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.976523] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] use_cow_images = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.976691] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] use_eventlog = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.976851] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] use_journal = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.977017] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] use_json = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.977184] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] use_rootwrap_daemon = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.977345] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] use_stderr = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.977534] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] use_syslog = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.977699] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vcpu_pin_set = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.977869] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plugging_is_fatal = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.978046] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plugging_timeout = 300 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.978218] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] virt_mkfs = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.978383] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] volume_usage_poll_interval = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.978550] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] watch_log_file = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.978716] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] web = /usr/share/spice-html5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 537.978902] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_concurrency.disable_process_locking = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.979460] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.979668] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.979895] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.980104] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.980292] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.980473] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.980660] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.auth_strategy = keystone {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.980828] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.compute_link_prefix = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.981014] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.981203] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.dhcp_domain = novalocal {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.981371] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.enable_instance_password = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.981539] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.glance_link_prefix = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.981709] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.981893] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.982073] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.instance_list_per_project_cells = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.982245] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.list_records_by_skipping_down_cells = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.982413] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.local_metadata_per_cell = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.982591] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.max_limit = 1000 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.982760] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.metadata_cache_expiration = 15 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.982935] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.neutron_default_tenant_id = default {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.983123] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.response_validation = warn {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.983299] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.use_neutron_default_nets = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.983472] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.983641] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.983817] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.983996] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.984181] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.vendordata_dynamic_targets = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.984351] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.vendordata_jsonfile_path = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.984536] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.984732] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.backend = dogpile.cache.memcached {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.984904] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.backend_argument = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.985094] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.config_prefix = cache.oslo {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.985339] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.dead_timeout = 60.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.985549] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.debug_cache_backend = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.985728] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.enable_retry_client = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.985897] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.enable_socket_keepalive = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.986085] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.enabled = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.986330] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.enforce_fips_mode = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.986520] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.expiration_time = 600 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.986697] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.hashclient_retry_attempts = 2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.986867] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.hashclient_retry_delay = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.987043] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.memcache_dead_retry = 300 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.987214] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.memcache_password = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.987386] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.987576] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.987750] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.memcache_pool_maxsize = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.987919] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.988101] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.memcache_sasl_enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.988284] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.988455] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.memcache_socket_timeout = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.988619] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.memcache_username = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.988789] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.proxies = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.988957] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.redis_db = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.989134] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.redis_password = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.989309] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.redis_sentinel_service_name = mymaster {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.989490] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.989663] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.redis_server = localhost:6379 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.989858] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.redis_socket_timeout = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.990039] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.redis_username = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.990212] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.retry_attempts = 2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.990379] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.retry_delay = 0.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.990543] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.socket_keepalive_count = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.990706] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.socket_keepalive_idle = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.990872] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.socket_keepalive_interval = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.991039] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.tls_allowed_ciphers = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.991213] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.tls_cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.991369] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.tls_certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.991533] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.tls_enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.991696] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cache.tls_keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.991870] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.auth_section = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.992060] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.auth_type = password {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.992228] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.992408] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.catalog_info = volumev3::publicURL {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.992572] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.992739] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.992904] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.cross_az_attach = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.993081] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.debug = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.993245] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.endpoint_template = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.993413] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.http_retries = 3 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.993579] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.993739] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.993910] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.os_region_name = RegionOne {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.994085] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.994251] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cinder.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.994425] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.994590] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.cpu_dedicated_set = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.994749] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.cpu_shared_set = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.994916] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.image_type_exclude_list = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.995096] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.995265] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.max_concurrent_disk_ops = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.995429] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.max_disk_devices_to_attach = -1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.995593] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.995762] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.995927] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.resource_provider_association_refresh = 300 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.996102] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.996269] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.shutdown_retry_interval = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.996539] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.996731] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] conductor.workers = 2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.996915] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] console.allowed_origins = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.997093] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] console.ssl_ciphers = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.997272] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] console.ssl_minimum_version = default {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.997470] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] consoleauth.enforce_session_timeout = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.997658] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] consoleauth.token_ttl = 600 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.997835] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.997998] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.998178] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.998343] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.connect_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.998532] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.connect_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.998699] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.endpoint_override = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.998864] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.999037] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.999207] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.max_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.999366] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.min_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.999525] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.region_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.999685] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.retriable_status_codes = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 537.999875] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.service_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.000064] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.service_type = accelerator {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.000232] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.000396] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.status_code_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.000558] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.status_code_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.000719] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.000902] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.001078] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] cyborg.version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.001266] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.backend = sqlalchemy {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.001440] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.connection = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.001616] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.connection_debug = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.001782] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.connection_parameters = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.001951] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.connection_recycle_time = 3600 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.002126] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.connection_trace = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.002290] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.db_inc_retry_interval = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.002453] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.db_max_retries = 20 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.002617] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.db_max_retry_interval = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.002782] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.db_retry_interval = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.002945] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.max_overflow = 50 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.003122] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.max_pool_size = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.003287] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.max_retries = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.003455] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.003616] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.mysql_wsrep_sync_wait = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.003773] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.pool_timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.003933] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.retry_interval = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.004102] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.slave_connection = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.004269] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.sqlite_synchronous = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.004429] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] database.use_db_reconnect = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.004612] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.backend = sqlalchemy {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.004781] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.connection = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.004947] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.connection_debug = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.005130] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.connection_parameters = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.005298] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.connection_recycle_time = 3600 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.005461] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.connection_trace = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.005625] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.db_inc_retry_interval = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.005789] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.db_max_retries = 20 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.005950] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.db_max_retry_interval = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.006125] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.db_retry_interval = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.006288] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.max_overflow = 50 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.006522] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.max_pool_size = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.006708] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.max_retries = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.006884] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.007059] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.007228] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.pool_timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.007423] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.retry_interval = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.007591] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.slave_connection = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.007761] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] api_database.sqlite_synchronous = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.007943] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] devices.enabled_mdev_types = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.008139] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.008316] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ephemeral_storage_encryption.default_format = luks {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.008502] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ephemeral_storage_encryption.enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.008675] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.008850] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.api_servers = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.009025] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.009197] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.009395] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.009565] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.connect_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.009729] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.connect_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.009894] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.debug = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.010073] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.default_trusted_certificate_ids = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.010246] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.enable_certificate_validation = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.010409] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.enable_rbd_download = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.010570] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.endpoint_override = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.010771] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.010938] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.011060] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.max_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.011223] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.min_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.011388] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.num_retries = 3 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.011562] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.rbd_ceph_conf = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.011724] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.rbd_connect_timeout = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.011921] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.rbd_pool = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.012108] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.rbd_user = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.012275] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.region_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.012440] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.retriable_status_codes = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.012600] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.service_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.012768] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.service_type = image {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.012930] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.013100] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.status_code_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.013261] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.status_code_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.013421] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.013705] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.013775] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.verify_glance_signatures = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.013930] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] glance.version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.014112] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] guestfs.debug = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.014283] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] mks.enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.014634] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.014828] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] image_cache.manager_interval = 2400 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.015007] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] image_cache.precache_concurrency = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.015187] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] image_cache.remove_unused_base_images = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.015358] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.015530] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.015710] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] image_cache.subdirectory_name = _base {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.015892] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.api_max_retries = 60 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.016069] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.api_retry_interval = 2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.016237] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.auth_section = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.016403] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.auth_type = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.016644] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.016826] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.016997] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.017182] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.conductor_group = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.017345] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.connect_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.017540] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.connect_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.017715] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.endpoint_override = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.017881] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.018053] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.018218] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.max_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.018389] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.min_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.018588] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.peer_list = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.018756] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.region_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.018918] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.retriable_status_codes = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.019096] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.serial_console_state_timeout = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.019263] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.service_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.019435] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.service_type = baremetal {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.019597] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.shard = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.019788] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.019967] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.status_code_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.020144] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.status_code_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.020306] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.020489] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.020655] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ironic.version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.020868] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.021055] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] key_manager.fixed_key = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.021209] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.021373] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.barbican_api_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.021548] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.barbican_endpoint = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.021725] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.barbican_endpoint_type = public {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.021889] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.barbican_region_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.022061] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.022225] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.022391] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.022558] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.022719] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.022886] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.number_of_retries = 60 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.023065] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.retry_delay = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.023236] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.send_service_user_token = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.023399] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.023562] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.023724] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.verify_ssl = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.023882] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican.verify_ssl_path = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.024064] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican_service_user.auth_section = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.024234] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican_service_user.auth_type = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.024396] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican_service_user.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.024561] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican_service_user.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.024727] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican_service_user.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.024891] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican_service_user.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.025063] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican_service_user.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.025231] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican_service_user.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.025398] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] barbican_service_user.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.025565] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.approle_role_id = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.025728] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.approle_secret_id = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.025903] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.kv_mountpoint = secret {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.026078] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.kv_path = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.026249] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.kv_version = 2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.026414] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.namespace = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.026576] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.root_token_id = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.026830] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.ssl_ca_crt_file = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.027014] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.timeout = 60.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.027187] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.use_ssl = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.027359] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.027566] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.auth_section = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.027740] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.auth_type = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.027905] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.028083] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.028250] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.028431] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.connect_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.028606] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.connect_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.028769] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.endpoint_override = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.028936] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.029112] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.029276] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.max_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.029436] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.min_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.029602] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.region_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.029779] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.retriable_status_codes = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.029952] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.service_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.030138] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.service_type = identity {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.030304] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.030465] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.status_code_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.030629] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.status_code_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.030792] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.030971] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.031164] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] keystone.version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.031353] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.connection_uri = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.031518] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.cpu_mode = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.031687] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.cpu_model_extra_flags = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.031893] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.cpu_models = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.032124] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.cpu_power_governor_high = performance {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.032309] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.cpu_power_governor_low = powersave {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.032478] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.cpu_power_management = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.032656] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.032827] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.device_detach_attempts = 8 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.032994] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.device_detach_timeout = 20 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.033179] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.disk_cachemodes = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.033344] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.disk_prefix = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.033512] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.enabled_perf_events = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.033683] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.file_backed_memory = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.033850] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.gid_maps = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.034020] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.hw_disk_discard = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.034185] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.hw_machine_type = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.034359] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.images_rbd_ceph_conf = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.034529] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.034697] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.034871] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.images_rbd_glance_store_name = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.035057] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.images_rbd_pool = rbd {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.035233] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.images_type = default {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.035395] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.images_volume_group = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.035561] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.inject_key = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.035726] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.inject_partition = -2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.035890] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.inject_password = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.036067] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.iscsi_iface = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.036237] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.iser_use_multipath = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.036411] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_bandwidth = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.036579] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.036814] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_downtime = 500 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.037008] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.037183] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.037346] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_inbound_addr = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.037540] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.037714] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_permit_post_copy = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.037879] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_scheme = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.038066] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_timeout_action = abort {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.038242] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_tunnelled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.038423] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_uri = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.038604] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.live_migration_with_native_tls = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.038770] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.max_queues = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.038940] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.039188] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.039357] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.nfs_mount_options = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.039649] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.039855] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.040043] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.num_iser_scan_tries = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.040217] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.num_memory_encrypted_guests = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.040384] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.040552] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.num_pcie_ports = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.040718] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.num_volume_scan_tries = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.040887] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.pmem_namespaces = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.041063] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.quobyte_client_cfg = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.041342] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.041521] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.rbd_connect_timeout = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.041693] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.041860] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.042064] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.rbd_secret_uuid = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.042202] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.rbd_user = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.042371] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.042546] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.remote_filesystem_transport = ssh {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.042712] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.rescue_image_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.042877] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.rescue_kernel_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.043045] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.rescue_ramdisk_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.043220] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.043384] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.rx_queue_size = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.043555] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.smbfs_mount_options = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.043828] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.044018] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.snapshot_compression = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.044188] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.snapshot_image_format = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.044407] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.044579] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.sparse_logical_volumes = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.044747] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.swtpm_enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.044920] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.swtpm_group = tss {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.045103] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.swtpm_user = tss {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.045280] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.sysinfo_serial = unique {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.045445] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.tb_cache_size = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.045612] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.tx_queue_size = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.045780] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.uid_maps = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.045946] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.use_virtio_for_bridges = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.046131] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.virt_type = kvm {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.046305] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.volume_clear = zero {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.046474] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.volume_clear_size = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.046645] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.volume_use_multipath = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.046874] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.vzstorage_cache_path = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.047084] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.047264] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.vzstorage_mount_group = qemu {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.047460] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.vzstorage_mount_opts = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.047708] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.048015] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.048209] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.vzstorage_mount_user = stack {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.048387] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.048591] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.auth_section = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.048779] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.auth_type = password {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.048948] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.049126] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.049299] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.049464] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.connect_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.049628] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.connect_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.049828] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.default_floating_pool = public {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.050013] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.endpoint_override = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.050185] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.extension_sync_interval = 600 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.050352] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.http_retries = 3 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.050517] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.050684] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.050846] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.max_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.051031] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.051199] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.min_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.051375] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.ovs_bridge = br-int {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.051534] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.physnets = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.051709] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.region_name = RegionOne {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.051872] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.retriable_status_codes = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.052054] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.service_metadata_proxy = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.052222] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.service_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.052393] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.service_type = network {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.052556] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.052716] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.status_code_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.052875] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.status_code_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.053046] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.053230] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.053394] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] neutron.version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.053571] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] notifications.bdms_in_notifications = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.053750] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] notifications.default_level = INFO {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.053925] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] notifications.notification_format = unversioned {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.054108] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] notifications.notify_on_state_change = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.054292] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.054470] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] pci.alias = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.054647] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] pci.device_spec = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.054814] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] pci.report_in_placement = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.054992] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.auth_section = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.055179] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.auth_type = password {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.055349] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.055513] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.055675] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.055838] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.055999] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.connect_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.056176] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.connect_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.056335] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.default_domain_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.056497] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.default_domain_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.056658] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.domain_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.056816] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.domain_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.057066] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.endpoint_override = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.057247] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.057443] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.057613] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.max_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.057780] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.min_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.057954] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.password = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.058131] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.project_domain_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.058303] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.project_domain_name = Default {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.058474] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.project_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.058651] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.project_name = service {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.058821] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.region_name = RegionOne {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.058987] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.retriable_status_codes = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.059163] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.service_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.059335] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.service_type = placement {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.059498] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.059664] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.status_code_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.059859] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.status_code_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.060087] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.system_scope = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.060208] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.060369] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.trust_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.060531] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.user_domain_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.060706] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.user_domain_name = Default {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.060865] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.user_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.061057] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.username = nova {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.061270] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.061412] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] placement.version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.061598] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.cores = 20 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.061768] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.count_usage_from_placement = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.061943] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.062134] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.injected_file_content_bytes = 10240 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.062306] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.injected_file_path_length = 255 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.062479] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.injected_files = 5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.062651] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.instances = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.062822] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.key_pairs = 100 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.062991] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.metadata_items = 128 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.063175] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.ram = 51200 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.063344] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.recheck_quota = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.063512] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.server_group_members = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.063684] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] quota.server_groups = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.063865] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.064042] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.064211] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler.image_metadata_prefilter = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.064378] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.064545] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler.max_attempts = 3 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.064711] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler.max_placement_results = 1000 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.064876] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.065048] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler.query_placement_for_image_type_support = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.065215] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.065425] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] scheduler.workers = 2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.065614] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.065790] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.065974] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.066165] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.066337] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.066504] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.066672] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.066862] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.067123] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.host_subset_size = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.067318] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.067514] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.067691] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.067863] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.isolated_hosts = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.068046] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.isolated_images = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.068220] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.068389] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.068559] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.068730] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.pci_in_placement = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.068897] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.069077] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.069247] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.069414] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.069584] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.069751] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.069915] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.track_instance_changes = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.070110] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.070288] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] metrics.required = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.070457] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] metrics.weight_multiplier = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.070624] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.070792] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] metrics.weight_setting = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.071121] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.071303] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] serial_console.enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.071486] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] serial_console.port_range = 10000:20000 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.071664] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.071835] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.072023] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] serial_console.serialproxy_port = 6083 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.072194] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_user.auth_section = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.072393] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_user.auth_type = password {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.072531] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_user.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.072694] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_user.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.072858] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_user.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.073065] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_user.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.073240] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_user.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.073418] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_user.send_service_user_token = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.073589] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_user.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.073754] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] service_user.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.073929] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.agent_enabled = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.074110] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.074441] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.074642] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.074818] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.html5proxy_port = 6082 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.074987] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.image_compression = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.075162] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.jpeg_compression = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.075357] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.playback_compression = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.075566] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.require_secure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.075754] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.server_listen = 127.0.0.1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.075928] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.076107] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.streaming_mode = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.076274] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] spice.zlib_compression = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.076446] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] upgrade_levels.baseapi = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.076627] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] upgrade_levels.compute = auto {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.076794] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] upgrade_levels.conductor = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.076952] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] upgrade_levels.scheduler = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.077210] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vendordata_dynamic_auth.auth_section = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.077430] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vendordata_dynamic_auth.auth_type = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.077586] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vendordata_dynamic_auth.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.077758] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vendordata_dynamic_auth.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.077926] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.078102] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vendordata_dynamic_auth.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.078267] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vendordata_dynamic_auth.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.078434] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.078598] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vendordata_dynamic_auth.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.078778] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.api_retry_count = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.078941] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.ca_file = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.079129] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.cache_prefix = devstack-image-cache {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.079301] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.cluster_name = testcl1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.079470] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.connection_pool_size = 10 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.079634] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.console_delay_seconds = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.079801] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.datastore_regex = ^datastore.* {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.080017] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.080199] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.host_password = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.080370] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.host_port = 443 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.080542] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.host_username = administrator@vsphere.local {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.080717] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.insecure = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.080884] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.integration_bridge = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.081066] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.maximum_objects = 100 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.081234] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.pbm_default_policy = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.081397] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.pbm_enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.081557] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.pbm_wsdl_location = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.081723] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.081884] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.serial_port_proxy_uri = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.082052] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.serial_port_service_uri = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.082225] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.task_poll_interval = 0.5 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.082400] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.use_linked_clone = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.082573] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.vnc_keymap = en-us {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.082739] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.vnc_port = 5900 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.082903] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vmware.vnc_port_total = 10000 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.083100] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vnc.auth_schemes = ['none'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.083282] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vnc.enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.083588] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.083780] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.083952] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vnc.novncproxy_port = 6080 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.084149] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vnc.server_listen = 127.0.0.1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.084325] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.084490] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vnc.vencrypt_ca_certs = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.084652] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vnc.vencrypt_client_cert = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.084815] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vnc.vencrypt_client_key = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.085032] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.085224] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.disable_deep_image_inspection = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.085395] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.085562] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.085728] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.085893] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.disable_rootwrap = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.086069] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.enable_numa_live_migration = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.086235] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.086398] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.086564] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.086732] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.libvirt_disable_apic = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.086900] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.087075] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.087331] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.087556] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.087745] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.087917] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.088094] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.088261] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.088425] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.088622] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.088819] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.088993] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.client_socket_timeout = 900 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.089183] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.default_pool_size = 1000 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.089356] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.keep_alive = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.089525] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.max_header_line = 16384 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.089694] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.secure_proxy_ssl_header = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.089857] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.ssl_ca_file = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.090027] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.ssl_cert_file = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.090193] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.ssl_key_file = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.090359] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.tcp_keepidle = 600 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.090537] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.090712] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] zvm.ca_file = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.090876] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] zvm.cloud_connector_url = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.091191] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.091369] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] zvm.reachable_timeout = 300 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.091565] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_policy.enforce_new_defaults = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.091944] env[61985]: WARNING oslo_config.cfg [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 538.092152] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_policy.enforce_scope = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.092334] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_policy.policy_default_rule = default {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.092522] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.092704] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_policy.policy_file = policy.yaml {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.092885] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.093061] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.093229] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.093390] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.093554] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.093721] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.093898] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.094087] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.connection_string = messaging:// {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.094261] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.enabled = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.094436] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.es_doc_type = notification {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.094607] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.es_scroll_size = 10000 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.094815] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.es_scroll_time = 2m {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.094992] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.filter_error_trace = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.095184] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.hmac_keys = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.095354] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.sentinel_service_name = mymaster {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.095561] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.socket_timeout = 0.1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.095736] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.trace_requests = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.095900] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler.trace_sqlalchemy = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.096103] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler_jaeger.process_tags = {} {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.096277] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler_jaeger.service_name_prefix = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.096442] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] profiler_otlp.service_name_prefix = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.096612] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] remote_debug.host = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.096774] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] remote_debug.port = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.096988] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.097183] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.097474] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.097654] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.097825] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.097992] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.098172] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.098340] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.098507] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.098687] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.098852] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.099038] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.099221] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.099391] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.099565] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.099735] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.099901] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.100085] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.100256] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.100420] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.100591] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.100756] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.100921] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.101099] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.101268] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.101434] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.101663] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.101767] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.101947] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.102113] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.ssl = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.102290] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.102466] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.102635] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.102809] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.102986] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.ssl_version = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.103164] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.103359] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.103530] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_notifications.retry = -1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.103725] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.103932] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_messaging_notifications.transport_url = **** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.104126] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.auth_section = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.104296] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.auth_type = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.104461] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.cafile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.104625] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.certfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.104789] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.collect_timing = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.104949] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.connect_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.105124] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.connect_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.105285] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.endpoint_id = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.105443] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.endpoint_override = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.105606] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.insecure = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.105765] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.keyfile = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.105922] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.max_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.106093] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.min_version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.106254] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.region_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.106418] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.retriable_status_codes = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.106580] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.service_name = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.106769] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.service_type = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.106952] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.split_loggers = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.107133] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.status_code_retries = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.107295] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.status_code_retry_delay = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.107573] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.timeout = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.107745] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.valid_interfaces = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.107909] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_limit.version = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.108092] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_reports.file_event_handler = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.108267] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.108429] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] oslo_reports.log_dir = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.108606] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.108768] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.108930] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.109112] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.109282] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.109446] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.109622] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.109787] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_ovs_privileged.group = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.109949] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.110132] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.110303] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.110463] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] vif_plug_ovs_privileged.user = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.110639] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_linux_bridge.flat_interface = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.110822] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.110999] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.111191] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.111365] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.111535] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.111711] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.111867] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.112073] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.112235] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_ovs.isolate_vif = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.112404] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.112572] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.112744] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.112927] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_ovs.ovsdb_interface = native {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.113089] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_vif_ovs.per_port_bridge = False {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.113263] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_brick.lock_path = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.113430] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.113660] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.113854] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] privsep_osbrick.capabilities = [21] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.114030] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] privsep_osbrick.group = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.114199] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] privsep_osbrick.helper_command = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.114374] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.114536] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.114699] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] privsep_osbrick.user = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.114875] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.115048] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] nova_sys_admin.group = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.115216] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] nova_sys_admin.helper_command = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.115386] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.115554] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.115719] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] nova_sys_admin.user = None {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 538.115851] env[61985]: DEBUG oslo_service.service [None req-6bfd85a5-e577-4314-8571-45ef6a3be50c None None] ******************************************************************************** {{(pid=61985) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 538.116360] env[61985]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 538.620122] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Getting list of instances from cluster (obj){ [ 538.620122] env[61985]: value = "domain-c8" [ 538.620122] env[61985]: _type = "ClusterComputeResource" [ 538.620122] env[61985]: } {{(pid=61985) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 538.621319] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74aea50-494d-4616-a932-2bb5aeb5984e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.630267] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Got total of 0 instances {{(pid=61985) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 538.630858] env[61985]: WARNING nova.virt.vmwareapi.driver [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 538.631354] env[61985]: INFO nova.virt.node [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Generated node identity aed7e5b3-c662-4538-8447-c4f67b460215 [ 538.631595] env[61985]: INFO nova.virt.node [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Wrote node identity aed7e5b3-c662-4538-8447-c4f67b460215 to /opt/stack/data/n-cpu-1/compute_id [ 539.134722] env[61985]: WARNING nova.compute.manager [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Compute nodes ['aed7e5b3-c662-4538-8447-c4f67b460215'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 540.140490] env[61985]: INFO nova.compute.manager [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 541.150020] env[61985]: WARNING nova.compute.manager [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 541.150020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.150020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.150020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.150020] env[61985]: DEBUG nova.compute.resource_tracker [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 541.150020] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6219d54-67af-4f70-8bda-eef095aff51e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.157870] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb5973f-1660-4f3c-a818-e0ea915fd0af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.172411] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1ea456-19b8-4d59-b63f-7bce7c19de7f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.178463] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47857e99-a39c-468e-9f8c-ce65eaa76cba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.207856] env[61985]: DEBUG nova.compute.resource_tracker [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181624MB free_disk=44GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 541.209129] env[61985]: DEBUG oslo_concurrency.lockutils [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.209129] env[61985]: DEBUG oslo_concurrency.lockutils [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.712779] env[61985]: WARNING nova.compute.resource_tracker [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] No compute node record for cpu-1:aed7e5b3-c662-4538-8447-c4f67b460215: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host aed7e5b3-c662-4538-8447-c4f67b460215 could not be found. [ 542.219674] env[61985]: INFO nova.compute.resource_tracker [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: aed7e5b3-c662-4538-8447-c4f67b460215 [ 543.727844] env[61985]: DEBUG nova.compute.resource_tracker [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 543.728152] env[61985]: DEBUG nova.compute.resource_tracker [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 543.905920] env[61985]: INFO nova.scheduler.client.report [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] [req-b6d5dfe8-a934-4dc5-ae62-81255fcc754e] Created resource provider record via placement API for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 543.921651] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1bbf864-c725-4923-aeed-213e689ac717 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.929424] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970f2573-44bd-4fae-8515-4b9f614001fb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.959462] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e32ce1c-8dea-4989-b7c8-74de23af3216 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.966030] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e65ad00-f4da-48d8-b3cf-25d11b131159 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.978719] env[61985]: DEBUG nova.compute.provider_tree [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 544.513207] env[61985]: DEBUG nova.scheduler.client.report [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 544.513449] env[61985]: DEBUG nova.compute.provider_tree [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 0 to 1 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 544.513589] env[61985]: DEBUG nova.compute.provider_tree [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 544.562480] env[61985]: DEBUG nova.compute.provider_tree [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 1 to 2 during operation: update_traits {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 545.067164] env[61985]: DEBUG nova.compute.resource_tracker [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 545.067685] env[61985]: DEBUG oslo_concurrency.lockutils [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.859s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.067685] env[61985]: DEBUG nova.service [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Creating RPC server for service compute {{(pid=61985) start /opt/stack/nova/nova/service.py:186}} [ 545.080325] env[61985]: DEBUG nova.service [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] Join ServiceGroup membership for this service compute {{(pid=61985) start /opt/stack/nova/nova/service.py:203}} [ 545.080499] env[61985]: DEBUG nova.servicegroup.drivers.db [None req-61c9dfad-4313-4ace-b6d4-509e4ac07609 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61985) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 548.081711] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.585432] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Getting list of instances from cluster (obj){ [ 548.585432] env[61985]: value = "domain-c8" [ 548.585432] env[61985]: _type = "ClusterComputeResource" [ 548.585432] env[61985]: } {{(pid=61985) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 548.586927] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493756dc-d229-4fbe-bf80-89120a236516 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.595482] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Got total of 0 instances {{(pid=61985) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 548.595719] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 548.596038] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Getting list of instances from cluster (obj){ [ 548.596038] env[61985]: value = "domain-c8" [ 548.596038] env[61985]: _type = "ClusterComputeResource" [ 548.596038] env[61985]: } {{(pid=61985) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 548.596883] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3036a105-b9c0-4de2-b389-e85aeaa38024 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.604031] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Got total of 0 instances {{(pid=61985) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 583.374781] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Acquiring lock "51fe587c-e4a1-4eab-9036-bdb59355d788" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.376203] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Lock "51fe587c-e4a1-4eab-9036-bdb59355d788" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.881425] env[61985]: DEBUG nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 584.427161] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.427495] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.429889] env[61985]: INFO nova.compute.claims [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 584.727032] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquiring lock "1cec2750-5d68-497c-8155-22afe8725809" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.727032] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "1cec2750-5d68-497c-8155-22afe8725809" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.166582] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Acquiring lock "1719bf7e-1f12-4f40-a56e-60d179790dce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.166582] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Lock "1719bf7e-1f12-4f40-a56e-60d179790dce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.234477] env[61985]: DEBUG nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 585.545143] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe83fc8f-931e-4346-ac63-7baf615ca584 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.553304] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e96617-cfe4-4a8d-8884-99251abe79c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.588063] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d3066a-494a-42bb-9194-8ecc70b69890 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.601232] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8479eb65-c0d4-4d24-956c-b7db6b3d080f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.620574] env[61985]: DEBUG nova.compute.provider_tree [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.669030] env[61985]: DEBUG nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 585.764722] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.782877] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Acquiring lock "d5ea62be-7be5-4c9f-990c-5d0ccd768556" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.783294] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Lock "d5ea62be-7be5-4c9f-990c-5d0ccd768556" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.125322] env[61985]: DEBUG nova.scheduler.client.report [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.194962] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.290627] env[61985]: DEBUG nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 586.631213] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.204s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.632882] env[61985]: DEBUG nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 586.638026] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.873s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.640208] env[61985]: INFO nova.compute.claims [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 586.820174] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Acquiring lock "896ea04a-ad5b-40bb-b521-9788736012bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.820307] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Lock "896ea04a-ad5b-40bb-b521-9788736012bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.823728] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.139423] env[61985]: DEBUG nova.compute.utils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 587.142354] env[61985]: DEBUG nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 587.142872] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 587.323293] env[61985]: DEBUG nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.651250] env[61985]: DEBUG nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 587.808298] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5144bcc6-ea84-4fc7-88b3-863240756bc1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.817393] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c0bd75-554f-4dd3-afe9-771ff33b3e9e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.858836] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47595298-7757-44e1-96bc-c14fff4777f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.867574] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1be05e7-c8a2-4d52-a1f4-0a737c61a3b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.883149] env[61985]: DEBUG nova.compute.provider_tree [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.884895] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.997981] env[61985]: DEBUG nova.policy [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a222fb83d1db485f9ee70086ad8bd23c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '951929b195f848e5be75d7c04516e332', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 588.387839] env[61985]: DEBUG nova.scheduler.client.report [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.531045] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Successfully created port: 24c82e2f-b9fd-4f2b-9734-836530238eef {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.665943] env[61985]: DEBUG nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 588.766534] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 588.766534] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 588.766534] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 588.766758] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 588.766758] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 588.766817] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 588.766995] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 588.769594] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 588.770072] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 588.770261] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 588.770442] env[61985]: DEBUG nova.virt.hardware [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 588.771490] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a795a6ca-e996-4767-aaf6-ef05970cce5d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.787083] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54cbb64-c5b3-41b1-8efa-3a3c1e06211f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.805664] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06af3669-acc2-4ac3-95da-e36512712e55 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.895524] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.257s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.896129] env[61985]: DEBUG nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 588.899153] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.704s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.902700] env[61985]: INFO nova.compute.claims [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.408539] env[61985]: DEBUG nova.compute.utils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 589.410095] env[61985]: DEBUG nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 589.410403] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 589.600222] env[61985]: DEBUG nova.policy [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16c90171651d4790a1f6f706b29f4b90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97a294a9bd8444c480ac4762c1cc5312', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 589.918138] env[61985]: DEBUG nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 590.031683] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e69e11-3af9-4130-8cb3-e348ab07b599 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.040294] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9ddde7-9960-4bda-9a9d-aaeed4aba32b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.072281] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15547f6-fa2f-4c52-9411-aee961a7c5fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.080322] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d92b7d-21ee-400f-b633-1ebaa5b19161 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.094378] env[61985]: DEBUG nova.compute.provider_tree [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.599651] env[61985]: DEBUG nova.scheduler.client.report [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.767616] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Successfully created port: e0c1bd46-c8f9-40bf-9baf-c9224e158573 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.931166] env[61985]: DEBUG nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 590.953080] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.953333] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.953490] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.953670] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.953814] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.953958] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.954176] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.954331] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.954540] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.954652] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.954822] env[61985]: DEBUG nova.virt.hardware [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.955740] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64db70ff-0989-4aec-9c47-8acca8e56ef8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.964156] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082456f7-e24d-4773-84a7-daf9669f815c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.107035] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.204s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.107035] env[61985]: DEBUG nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 591.107035] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.284s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.109975] env[61985]: INFO nova.compute.claims [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 591.615432] env[61985]: DEBUG nova.compute.utils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.616815] env[61985]: DEBUG nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.616990] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 591.996047] env[61985]: DEBUG nova.policy [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0728c301040b4df89563149ef784c91b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22ed123cf9534ffaba6f5b6e2ea851f3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 592.133947] env[61985]: DEBUG nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 592.251208] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de2f5e8-a698-4952-a233-039091f0e967 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.261083] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5114a8a-9d70-4021-a192-b57557cfcdb8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.296677] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f97e38c-e6a5-4e2b-acc9-d56c8a09c04d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.305796] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4e2d19-67b1-481b-b146-cc85dc3d520f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.323195] env[61985]: DEBUG nova.compute.provider_tree [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.803098] env[61985]: ERROR nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 24c82e2f-b9fd-4f2b-9734-836530238eef, please check neutron logs for more information. [ 592.803098] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 592.803098] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.803098] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 592.803098] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.803098] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 592.803098] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.803098] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 592.803098] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.803098] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 592.803098] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.803098] env[61985]: ERROR nova.compute.manager raise self.value [ 592.803098] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.803098] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 592.803098] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.803098] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 592.804167] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.804167] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 592.804167] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 24c82e2f-b9fd-4f2b-9734-836530238eef, please check neutron logs for more information. [ 592.804167] env[61985]: ERROR nova.compute.manager [ 592.804167] env[61985]: Traceback (most recent call last): [ 592.804167] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 592.804167] env[61985]: listener.cb(fileno) [ 592.804167] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.804167] env[61985]: result = function(*args, **kwargs) [ 592.804167] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.804167] env[61985]: return func(*args, **kwargs) [ 592.804167] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.804167] env[61985]: raise e [ 592.804167] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.804167] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 592.804167] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.804167] env[61985]: created_port_ids = self._update_ports_for_instance( [ 592.804167] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.804167] env[61985]: with excutils.save_and_reraise_exception(): [ 592.804167] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.804167] env[61985]: self.force_reraise() [ 592.804167] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.804167] env[61985]: raise self.value [ 592.804167] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.804167] env[61985]: updated_port = self._update_port( [ 592.804167] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.804167] env[61985]: _ensure_no_port_binding_failure(port) [ 592.804167] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.804167] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 592.805606] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 24c82e2f-b9fd-4f2b-9734-836530238eef, please check neutron logs for more information. [ 592.805606] env[61985]: Removing descriptor: 14 [ 592.805606] env[61985]: ERROR nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 24c82e2f-b9fd-4f2b-9734-836530238eef, please check neutron logs for more information. [ 592.805606] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Traceback (most recent call last): [ 592.805606] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 592.805606] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] yield resources [ 592.805606] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 592.805606] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] self.driver.spawn(context, instance, image_meta, [ 592.805606] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 592.805606] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.805606] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.805606] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] vm_ref = self.build_virtual_machine(instance, [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] for vif in network_info: [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] return self._sync_wrapper(fn, *args, **kwargs) [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] self.wait() [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] self[:] = self._gt.wait() [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] return self._exit_event.wait() [ 592.806067] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] result = hub.switch() [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] return self.greenlet.switch() [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] result = function(*args, **kwargs) [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] return func(*args, **kwargs) [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] raise e [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] nwinfo = self.network_api.allocate_for_instance( [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.806519] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] created_port_ids = self._update_ports_for_instance( [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] with excutils.save_and_reraise_exception(): [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] self.force_reraise() [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] raise self.value [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] updated_port = self._update_port( [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] _ensure_no_port_binding_failure(port) [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.806874] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] raise exception.PortBindingFailed(port_id=port['id']) [ 592.807213] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] nova.exception.PortBindingFailed: Binding failed for port 24c82e2f-b9fd-4f2b-9734-836530238eef, please check neutron logs for more information. [ 592.807213] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] [ 592.807213] env[61985]: INFO nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Terminating instance [ 592.809217] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Acquiring lock "refresh_cache-51fe587c-e4a1-4eab-9036-bdb59355d788" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.809538] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Acquired lock "refresh_cache-51fe587c-e4a1-4eab-9036-bdb59355d788" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.809538] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 592.828937] env[61985]: DEBUG nova.scheduler.client.report [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.086816] env[61985]: DEBUG nova.compute.manager [req-9c626a31-8391-4186-b6a7-52a8a04e6ee8 req-350a79ff-a133-4d86-aa0c-65ddf3e0f8b4 service nova] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Received event network-changed-24c82e2f-b9fd-4f2b-9734-836530238eef {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 593.087076] env[61985]: DEBUG nova.compute.manager [req-9c626a31-8391-4186-b6a7-52a8a04e6ee8 req-350a79ff-a133-4d86-aa0c-65ddf3e0f8b4 service nova] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Refreshing instance network info cache due to event network-changed-24c82e2f-b9fd-4f2b-9734-836530238eef. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 593.087273] env[61985]: DEBUG oslo_concurrency.lockutils [req-9c626a31-8391-4186-b6a7-52a8a04e6ee8 req-350a79ff-a133-4d86-aa0c-65ddf3e0f8b4 service nova] Acquiring lock "refresh_cache-51fe587c-e4a1-4eab-9036-bdb59355d788" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.144837] env[61985]: DEBUG nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 593.175045] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 593.175045] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 593.175322] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 593.175322] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 593.175568] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 593.175659] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 593.175897] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 593.175956] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 593.176820] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 593.176820] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 593.176820] env[61985]: DEBUG nova.virt.hardware [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 593.180255] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056cd4ef-265b-4f18-a0de-afb6bacfa320 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.188283] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6534c5-e287-48d2-a430-64aba11e5cbe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.339017] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.232s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.339539] env[61985]: DEBUG nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 593.342197] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.457s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.343557] env[61985]: INFO nova.compute.claims [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.371503] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.499691] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 593.500037] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 593.500259] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 593.500354] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Rebuilding the list of instances to heal {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9936}} [ 593.610668] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.848814] env[61985]: DEBUG nova.compute.utils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.855838] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Successfully created port: 1a9802c4-5560-461c-9375-3d5a0fc3e1b7 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.857842] env[61985]: DEBUG nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 593.858032] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 594.007860] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 594.007860] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 594.007860] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 594.007860] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 594.007860] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 594.007860] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Didn't find any instances for network info cache update. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10018}} [ 594.008139] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 594.008533] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 594.008867] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 594.009436] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 594.009770] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 594.011422] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 594.011422] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 594.011422] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 594.091415] env[61985]: DEBUG nova.policy [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '39b3f6027b694caf9ea218c10024f1ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7110c782f793489daec744b11f4f065d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 594.114252] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Releasing lock "refresh_cache-51fe587c-e4a1-4eab-9036-bdb59355d788" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.114399] env[61985]: DEBUG nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 594.114603] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 594.114912] env[61985]: DEBUG oslo_concurrency.lockutils [req-9c626a31-8391-4186-b6a7-52a8a04e6ee8 req-350a79ff-a133-4d86-aa0c-65ddf3e0f8b4 service nova] Acquired lock "refresh_cache-51fe587c-e4a1-4eab-9036-bdb59355d788" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.115092] env[61985]: DEBUG nova.network.neutron [req-9c626a31-8391-4186-b6a7-52a8a04e6ee8 req-350a79ff-a133-4d86-aa0c-65ddf3e0f8b4 service nova] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Refreshing network info cache for port 24c82e2f-b9fd-4f2b-9734-836530238eef {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 594.116263] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6ce3173-dd0a-4b16-80da-0fc284a45c37 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.128663] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddd3914-9064-45f2-a745-bf73cc6fe7df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.156926] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 51fe587c-e4a1-4eab-9036-bdb59355d788 could not be found. [ 594.157245] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 594.157773] env[61985]: INFO nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Took 0.04 seconds to destroy the instance on the hypervisor. [ 594.158103] env[61985]: DEBUG oslo.service.loopingcall [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 594.158759] env[61985]: DEBUG nova.compute.manager [-] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.158759] env[61985]: DEBUG nova.network.neutron [-] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 594.248171] env[61985]: DEBUG nova.network.neutron [-] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.352861] env[61985]: DEBUG nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 594.479112] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484bba8d-05a5-493d-a262-e43c25438916 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.491336] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bbfdd4-77ac-4116-8fdc-40a59638b4a0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.529949] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.531072] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9df362-58d4-423f-bfdf-23e9bb96bfb7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.541918] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73d7799-dd81-498b-9904-0d662c58351f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.556306] env[61985]: DEBUG nova.compute.provider_tree [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.687837] env[61985]: DEBUG nova.network.neutron [req-9c626a31-8391-4186-b6a7-52a8a04e6ee8 req-350a79ff-a133-4d86-aa0c-65ddf3e0f8b4 service nova] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.755926] env[61985]: DEBUG nova.network.neutron [-] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.060492] env[61985]: DEBUG nova.scheduler.client.report [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.118722] env[61985]: DEBUG nova.network.neutron [req-9c626a31-8391-4186-b6a7-52a8a04e6ee8 req-350a79ff-a133-4d86-aa0c-65ddf3e0f8b4 service nova] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.260709] env[61985]: INFO nova.compute.manager [-] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Took 1.10 seconds to deallocate network for instance. [ 595.266787] env[61985]: DEBUG nova.compute.claims [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 595.267085] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.364445] env[61985]: DEBUG nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 595.405219] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 595.405219] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 595.405219] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.405827] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 595.405827] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.406395] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 595.406835] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 595.407168] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 595.407530] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 595.407864] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 595.408158] env[61985]: DEBUG nova.virt.hardware [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 595.410032] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d24508-4f22-496e-9614-0af7a320ccc0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.419219] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d83e66-f0cf-46f0-83f0-ad3b39bb63b6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.565760] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.223s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.566323] env[61985]: DEBUG nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 595.574533] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.044s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.574533] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.574533] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 595.574966] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 0.308s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.582077] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf2e2f7-c52e-4cb8-87b4-1c8cae547021 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.595344] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bff38cd-ef7d-4ba5-8544-44b50a604623 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.619617] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cacb15c-1793-419d-b2c0-b3c4f3b57dcc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.622829] env[61985]: DEBUG oslo_concurrency.lockutils [req-9c626a31-8391-4186-b6a7-52a8a04e6ee8 req-350a79ff-a133-4d86-aa0c-65ddf3e0f8b4 service nova] Releasing lock "refresh_cache-51fe587c-e4a1-4eab-9036-bdb59355d788" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.633534] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92efd940-8e13-4ae2-94f1-d05fb7ba8982 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.671380] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181623MB free_disk=44GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 595.671380] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.994784] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Successfully created port: 43d12649-e1b7-42d1-a35e-35a76f24d507 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.083989] env[61985]: DEBUG nova.compute.utils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.088125] env[61985]: DEBUG nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.088125] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 596.251370] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc7349c0-6e3d-48f8-859f-6845de351f43 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.261777] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb4f577-32fc-4128-b0f9-cc67f414a9ce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.315874] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e97fdad-6ec5-4397-b602-1f675b4be8a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.323484] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e60b22b-7969-4622-b2f9-18aec0a6caf5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.341750] env[61985]: DEBUG nova.compute.provider_tree [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.533634] env[61985]: DEBUG nova.policy [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14a1aff90e8742308908c0f9c2178e22', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f6779234a0a4691bbbcde0c0e6e3f35', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.597452] env[61985]: DEBUG nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 596.844226] env[61985]: DEBUG nova.scheduler.client.report [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.036162] env[61985]: ERROR nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e0c1bd46-c8f9-40bf-9baf-c9224e158573, please check neutron logs for more information. [ 597.036162] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.036162] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.036162] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.036162] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.036162] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.036162] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.036162] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.036162] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.036162] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 597.036162] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.036162] env[61985]: ERROR nova.compute.manager raise self.value [ 597.036162] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.036162] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.036162] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.036162] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.036713] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.036713] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.036713] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e0c1bd46-c8f9-40bf-9baf-c9224e158573, please check neutron logs for more information. [ 597.036713] env[61985]: ERROR nova.compute.manager [ 597.036713] env[61985]: Traceback (most recent call last): [ 597.036713] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.036713] env[61985]: listener.cb(fileno) [ 597.036713] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.036713] env[61985]: result = function(*args, **kwargs) [ 597.036713] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.036713] env[61985]: return func(*args, **kwargs) [ 597.036713] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.036713] env[61985]: raise e [ 597.036713] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.036713] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 597.036713] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.036713] env[61985]: created_port_ids = self._update_ports_for_instance( [ 597.036713] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.036713] env[61985]: with excutils.save_and_reraise_exception(): [ 597.036713] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.036713] env[61985]: self.force_reraise() [ 597.036713] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.036713] env[61985]: raise self.value [ 597.036713] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.036713] env[61985]: updated_port = self._update_port( [ 597.036713] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.036713] env[61985]: _ensure_no_port_binding_failure(port) [ 597.036713] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.036713] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.037558] env[61985]: nova.exception.PortBindingFailed: Binding failed for port e0c1bd46-c8f9-40bf-9baf-c9224e158573, please check neutron logs for more information. [ 597.037558] env[61985]: Removing descriptor: 16 [ 597.037558] env[61985]: ERROR nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e0c1bd46-c8f9-40bf-9baf-c9224e158573, please check neutron logs for more information. [ 597.037558] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] Traceback (most recent call last): [ 597.037558] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 597.037558] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] yield resources [ 597.037558] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 597.037558] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] self.driver.spawn(context, instance, image_meta, [ 597.037558] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.037558] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.037558] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.037558] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] vm_ref = self.build_virtual_machine(instance, [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] for vif in network_info: [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] return self._sync_wrapper(fn, *args, **kwargs) [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] self.wait() [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] self[:] = self._gt.wait() [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] return self._exit_event.wait() [ 597.037906] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] result = hub.switch() [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] return self.greenlet.switch() [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] result = function(*args, **kwargs) [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] return func(*args, **kwargs) [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] raise e [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] nwinfo = self.network_api.allocate_for_instance( [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.038335] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] created_port_ids = self._update_ports_for_instance( [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] with excutils.save_and_reraise_exception(): [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] self.force_reraise() [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] raise self.value [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] updated_port = self._update_port( [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] _ensure_no_port_binding_failure(port) [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.038734] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] raise exception.PortBindingFailed(port_id=port['id']) [ 597.039095] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] nova.exception.PortBindingFailed: Binding failed for port e0c1bd46-c8f9-40bf-9baf-c9224e158573, please check neutron logs for more information. [ 597.039095] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] [ 597.039095] env[61985]: INFO nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Terminating instance [ 597.039636] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquiring lock "refresh_cache-1cec2750-5d68-497c-8155-22afe8725809" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.039794] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquired lock "refresh_cache-1cec2750-5d68-497c-8155-22afe8725809" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.040013] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.350206] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.775s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.350206] env[61985]: ERROR nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 24c82e2f-b9fd-4f2b-9734-836530238eef, please check neutron logs for more information. [ 597.350206] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Traceback (most recent call last): [ 597.350206] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 597.350206] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] self.driver.spawn(context, instance, image_meta, [ 597.350206] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.350206] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.350206] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.350206] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] vm_ref = self.build_virtual_machine(instance, [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] for vif in network_info: [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] return self._sync_wrapper(fn, *args, **kwargs) [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] self.wait() [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] self[:] = self._gt.wait() [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] return self._exit_event.wait() [ 597.350659] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] result = hub.switch() [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] return self.greenlet.switch() [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] result = function(*args, **kwargs) [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] return func(*args, **kwargs) [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] raise e [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] nwinfo = self.network_api.allocate_for_instance( [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.350986] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] created_port_ids = self._update_ports_for_instance( [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] with excutils.save_and_reraise_exception(): [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] self.force_reraise() [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] raise self.value [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] updated_port = self._update_port( [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] _ensure_no_port_binding_failure(port) [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.351330] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] raise exception.PortBindingFailed(port_id=port['id']) [ 597.351638] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] nova.exception.PortBindingFailed: Binding failed for port 24c82e2f-b9fd-4f2b-9734-836530238eef, please check neutron logs for more information. [ 597.351638] env[61985]: ERROR nova.compute.manager [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] [ 597.351638] env[61985]: DEBUG nova.compute.utils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Binding failed for port 24c82e2f-b9fd-4f2b-9734-836530238eef, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.354786] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.683s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.359843] env[61985]: DEBUG nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Build of instance 51fe587c-e4a1-4eab-9036-bdb59355d788 was re-scheduled: Binding failed for port 24c82e2f-b9fd-4f2b-9734-836530238eef, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.360499] env[61985]: DEBUG nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 597.361914] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Acquiring lock "refresh_cache-51fe587c-e4a1-4eab-9036-bdb59355d788" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.361914] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Acquired lock "refresh_cache-51fe587c-e4a1-4eab-9036-bdb59355d788" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.361914] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.599927] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.614915] env[61985]: DEBUG nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 597.651917] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.652149] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.652282] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.652522] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.652593] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.652732] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.652934] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.653129] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.653266] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.653421] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.653934] env[61985]: DEBUG nova.virt.hardware [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.655962] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15cbf23a-38b6-4ad0-b78a-d3bae30d0c59 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.674520] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0dfd500-98cb-4d7c-9b49-ebe344cbc07c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.882700] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.901461] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.051469] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Successfully created port: 42cde311-8e8e-4b9a-a5eb-55168112473b {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 598.067297] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.341568] env[61985]: ERROR nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 43d12649-e1b7-42d1-a35e-35a76f24d507, please check neutron logs for more information. [ 598.341568] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.341568] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.341568] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.341568] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.341568] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.341568] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.341568] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.341568] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.341568] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 598.341568] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.341568] env[61985]: ERROR nova.compute.manager raise self.value [ 598.341568] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.341568] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.341568] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.341568] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.341999] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.341999] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.341999] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 43d12649-e1b7-42d1-a35e-35a76f24d507, please check neutron logs for more information. [ 598.341999] env[61985]: ERROR nova.compute.manager [ 598.341999] env[61985]: Traceback (most recent call last): [ 598.341999] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.341999] env[61985]: listener.cb(fileno) [ 598.341999] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.341999] env[61985]: result = function(*args, **kwargs) [ 598.341999] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.341999] env[61985]: return func(*args, **kwargs) [ 598.341999] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.341999] env[61985]: raise e [ 598.341999] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.341999] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 598.341999] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.341999] env[61985]: created_port_ids = self._update_ports_for_instance( [ 598.341999] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.341999] env[61985]: with excutils.save_and_reraise_exception(): [ 598.341999] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.341999] env[61985]: self.force_reraise() [ 598.341999] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.341999] env[61985]: raise self.value [ 598.341999] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.341999] env[61985]: updated_port = self._update_port( [ 598.341999] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.341999] env[61985]: _ensure_no_port_binding_failure(port) [ 598.341999] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.341999] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 598.342925] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 43d12649-e1b7-42d1-a35e-35a76f24d507, please check neutron logs for more information. [ 598.342925] env[61985]: Removing descriptor: 14 [ 598.342925] env[61985]: ERROR nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 43d12649-e1b7-42d1-a35e-35a76f24d507, please check neutron logs for more information. [ 598.342925] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Traceback (most recent call last): [ 598.342925] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 598.342925] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] yield resources [ 598.342925] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 598.342925] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] self.driver.spawn(context, instance, image_meta, [ 598.342925] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 598.342925] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.342925] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.342925] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] vm_ref = self.build_virtual_machine(instance, [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] for vif in network_info: [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] return self._sync_wrapper(fn, *args, **kwargs) [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] self.wait() [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] self[:] = self._gt.wait() [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] return self._exit_event.wait() [ 598.343241] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] result = hub.switch() [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] return self.greenlet.switch() [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] result = function(*args, **kwargs) [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] return func(*args, **kwargs) [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] raise e [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] nwinfo = self.network_api.allocate_for_instance( [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.343565] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] created_port_ids = self._update_ports_for_instance( [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] with excutils.save_and_reraise_exception(): [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] self.force_reraise() [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] raise self.value [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] updated_port = self._update_port( [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] _ensure_no_port_binding_failure(port) [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.343931] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] raise exception.PortBindingFailed(port_id=port['id']) [ 598.344234] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] nova.exception.PortBindingFailed: Binding failed for port 43d12649-e1b7-42d1-a35e-35a76f24d507, please check neutron logs for more information. [ 598.344234] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] [ 598.344234] env[61985]: INFO nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Terminating instance [ 598.345550] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Acquiring lock "refresh_cache-d5ea62be-7be5-4c9f-990c-5d0ccd768556" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.345752] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Acquired lock "refresh_cache-d5ea62be-7be5-4c9f-990c-5d0ccd768556" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.345948] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.385628] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Releasing lock "refresh_cache-1cec2750-5d68-497c-8155-22afe8725809" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.386091] env[61985]: DEBUG nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 598.386756] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 598.389415] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37839c24-21e3-4a0b-8356-f1f6797f281b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.401160] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6298b2d-7d4f-4974-8944-00c1c371741c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.433507] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1cec2750-5d68-497c-8155-22afe8725809 could not be found. [ 598.433789] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 598.434461] env[61985]: INFO nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Took 0.05 seconds to destroy the instance on the hypervisor. [ 598.434998] env[61985]: DEBUG oslo.service.loopingcall [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.435242] env[61985]: DEBUG nova.compute.manager [-] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.435343] env[61985]: DEBUG nova.network.neutron [-] [instance: 1cec2750-5d68-497c-8155-22afe8725809] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.467161] env[61985]: DEBUG nova.network.neutron [-] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.570926] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Releasing lock "refresh_cache-51fe587c-e4a1-4eab-9036-bdb59355d788" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.571187] env[61985]: DEBUG nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 598.571364] env[61985]: DEBUG nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.571558] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.594195] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.874031] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.894625] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 51fe587c-e4a1-4eab-9036-bdb59355d788 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 598.894793] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 1cec2750-5d68-497c-8155-22afe8725809 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 598.894924] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 1719bf7e-1f12-4f40-a56e-60d179790dce actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 598.895061] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance d5ea62be-7be5-4c9f-990c-5d0ccd768556 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 598.895183] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 896ea04a-ad5b-40bb-b521-9788736012bc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 598.895377] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 598.895520] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 598.952024] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.972747] env[61985]: DEBUG nova.network.neutron [-] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.993052] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7635fdd-267a-4508-8c98-2b7656648936 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.003596] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56eda0b7-f41c-46c2-a114-7e674ac0af4b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.041928] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a88bcdf-a651-419d-bfa9-8123d0f8a182 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.050071] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44f9268-89bc-4d21-b8b5-a82695d6f5ce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.068185] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.097105] env[61985]: DEBUG nova.network.neutron [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.454719] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Releasing lock "refresh_cache-d5ea62be-7be5-4c9f-990c-5d0ccd768556" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.455317] env[61985]: DEBUG nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 599.459211] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 599.459579] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08123458-4b1d-42df-8897-ef09705c4368 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.470531] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b66edd-e4ae-4d31-8149-dc3f752b9e72 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.483558] env[61985]: INFO nova.compute.manager [-] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Took 1.05 seconds to deallocate network for instance. [ 599.486430] env[61985]: DEBUG nova.compute.claims [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.486629] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.497069] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d5ea62be-7be5-4c9f-990c-5d0ccd768556 could not be found. [ 599.497302] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 599.497535] env[61985]: INFO nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Took 0.04 seconds to destroy the instance on the hypervisor. [ 599.498014] env[61985]: DEBUG oslo.service.loopingcall [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.498242] env[61985]: DEBUG nova.compute.manager [-] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.498338] env[61985]: DEBUG nova.network.neutron [-] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.516896] env[61985]: DEBUG nova.network.neutron [-] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.572437] env[61985]: ERROR nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1a9802c4-5560-461c-9375-3d5a0fc3e1b7, please check neutron logs for more information. [ 599.572437] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 599.572437] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.572437] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 599.572437] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.572437] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 599.572437] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.572437] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 599.572437] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.572437] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 599.572437] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.572437] env[61985]: ERROR nova.compute.manager raise self.value [ 599.572437] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.572437] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 599.572437] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.572437] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 599.572904] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.572904] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 599.572904] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1a9802c4-5560-461c-9375-3d5a0fc3e1b7, please check neutron logs for more information. [ 599.572904] env[61985]: ERROR nova.compute.manager [ 599.572904] env[61985]: Traceback (most recent call last): [ 599.572904] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 599.572904] env[61985]: listener.cb(fileno) [ 599.572904] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.572904] env[61985]: result = function(*args, **kwargs) [ 599.572904] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.572904] env[61985]: return func(*args, **kwargs) [ 599.572904] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.572904] env[61985]: raise e [ 599.572904] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.572904] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 599.572904] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.572904] env[61985]: created_port_ids = self._update_ports_for_instance( [ 599.572904] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.572904] env[61985]: with excutils.save_and_reraise_exception(): [ 599.572904] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.572904] env[61985]: self.force_reraise() [ 599.572904] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.572904] env[61985]: raise self.value [ 599.572904] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.572904] env[61985]: updated_port = self._update_port( [ 599.572904] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.572904] env[61985]: _ensure_no_port_binding_failure(port) [ 599.572904] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.572904] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.573643] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 1a9802c4-5560-461c-9375-3d5a0fc3e1b7, please check neutron logs for more information. [ 599.573643] env[61985]: Removing descriptor: 17 [ 599.573643] env[61985]: ERROR nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1a9802c4-5560-461c-9375-3d5a0fc3e1b7, please check neutron logs for more information. [ 599.573643] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Traceback (most recent call last): [ 599.573643] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 599.573643] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] yield resources [ 599.573643] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 599.573643] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] self.driver.spawn(context, instance, image_meta, [ 599.573643] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.573643] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.573643] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.573643] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] vm_ref = self.build_virtual_machine(instance, [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] for vif in network_info: [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] return self._sync_wrapper(fn, *args, **kwargs) [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] self.wait() [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] self[:] = self._gt.wait() [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] return self._exit_event.wait() [ 599.573990] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] result = hub.switch() [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] return self.greenlet.switch() [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] result = function(*args, **kwargs) [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] return func(*args, **kwargs) [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] raise e [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] nwinfo = self.network_api.allocate_for_instance( [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.574481] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] created_port_ids = self._update_ports_for_instance( [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] with excutils.save_and_reraise_exception(): [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] self.force_reraise() [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] raise self.value [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] updated_port = self._update_port( [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] _ensure_no_port_binding_failure(port) [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.574907] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] raise exception.PortBindingFailed(port_id=port['id']) [ 599.575324] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] nova.exception.PortBindingFailed: Binding failed for port 1a9802c4-5560-461c-9375-3d5a0fc3e1b7, please check neutron logs for more information. [ 599.575324] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] [ 599.575324] env[61985]: INFO nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Terminating instance [ 599.576452] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.581049] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Acquiring lock "refresh_cache-1719bf7e-1f12-4f40-a56e-60d179790dce" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.581246] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Acquired lock "refresh_cache-1719bf7e-1f12-4f40-a56e-60d179790dce" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.581417] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 599.601053] env[61985]: INFO nova.compute.manager [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] [instance: 51fe587c-e4a1-4eab-9036-bdb59355d788] Took 1.03 seconds to deallocate network for instance. [ 600.019360] env[61985]: DEBUG nova.network.neutron [-] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.088296] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 600.088296] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.733s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.088296] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 0.601s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.119463] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.237233] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.523189] env[61985]: INFO nova.compute.manager [-] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Took 1.02 seconds to deallocate network for instance. [ 600.528034] env[61985]: DEBUG nova.compute.claims [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.528180] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.646205] env[61985]: INFO nova.scheduler.client.report [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Deleted allocations for instance 51fe587c-e4a1-4eab-9036-bdb59355d788 [ 600.700802] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5421481f-2f1f-4956-a53e-783658e9b20b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.712139] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce5ac54e-2836-4936-83fc-c1d5fab247a8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.750277] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Releasing lock "refresh_cache-1719bf7e-1f12-4f40-a56e-60d179790dce" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.750618] env[61985]: DEBUG nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 600.750944] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 600.751802] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-27498c69-b2b8-47be-98e8-781378ae4b5f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.754634] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b3b4c6-8fc6-4308-a54d-a1cd69d8c33e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.764787] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bcdc86e-d715-4609-94de-7dab260b0a52 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.788880] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1719bf7e-1f12-4f40-a56e-60d179790dce could not be found. [ 600.788987] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 600.789205] env[61985]: INFO nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Took 0.04 seconds to destroy the instance on the hypervisor. [ 600.789448] env[61985]: DEBUG oslo.service.loopingcall [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.791991] env[61985]: DEBUG nova.compute.manager [-] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.793085] env[61985]: DEBUG nova.network.neutron [-] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 600.796965] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42ec38c-d34a-4f79-8625-f5e3e91dda0b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.816156] env[61985]: DEBUG nova.compute.provider_tree [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.838330] env[61985]: DEBUG nova.network.neutron [-] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.155542] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1ce8124f-eeee-4940-b7d1-df95783a045b tempest-ServerDiagnosticsNegativeTest-1172244291 tempest-ServerDiagnosticsNegativeTest-1172244291-project-member] Lock "51fe587c-e4a1-4eab-9036-bdb59355d788" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.779s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.209828] env[61985]: ERROR nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 42cde311-8e8e-4b9a-a5eb-55168112473b, please check neutron logs for more information. [ 601.209828] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.209828] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.209828] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.209828] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.209828] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.209828] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.209828] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.209828] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.209828] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 601.209828] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.209828] env[61985]: ERROR nova.compute.manager raise self.value [ 601.209828] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.209828] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.209828] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.209828] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.210309] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.210309] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.210309] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 42cde311-8e8e-4b9a-a5eb-55168112473b, please check neutron logs for more information. [ 601.210309] env[61985]: ERROR nova.compute.manager [ 601.210309] env[61985]: Traceback (most recent call last): [ 601.210309] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.210309] env[61985]: listener.cb(fileno) [ 601.210309] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.210309] env[61985]: result = function(*args, **kwargs) [ 601.210309] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.210309] env[61985]: return func(*args, **kwargs) [ 601.210309] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.210309] env[61985]: raise e [ 601.210309] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.210309] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 601.210309] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.210309] env[61985]: created_port_ids = self._update_ports_for_instance( [ 601.210309] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.210309] env[61985]: with excutils.save_and_reraise_exception(): [ 601.210309] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.210309] env[61985]: self.force_reraise() [ 601.210309] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.210309] env[61985]: raise self.value [ 601.210309] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.210309] env[61985]: updated_port = self._update_port( [ 601.210309] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.210309] env[61985]: _ensure_no_port_binding_failure(port) [ 601.210309] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.210309] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.211026] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 42cde311-8e8e-4b9a-a5eb-55168112473b, please check neutron logs for more information. [ 601.211026] env[61985]: Removing descriptor: 18 [ 601.211026] env[61985]: ERROR nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 42cde311-8e8e-4b9a-a5eb-55168112473b, please check neutron logs for more information. [ 601.211026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Traceback (most recent call last): [ 601.211026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 601.211026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] yield resources [ 601.211026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 601.211026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] self.driver.spawn(context, instance, image_meta, [ 601.211026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 601.211026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.211026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.211026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] vm_ref = self.build_virtual_machine(instance, [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] for vif in network_info: [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] return self._sync_wrapper(fn, *args, **kwargs) [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] self.wait() [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] self[:] = self._gt.wait() [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] return self._exit_event.wait() [ 601.211360] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] result = hub.switch() [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] return self.greenlet.switch() [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] result = function(*args, **kwargs) [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] return func(*args, **kwargs) [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] raise e [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] nwinfo = self.network_api.allocate_for_instance( [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.211678] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] created_port_ids = self._update_ports_for_instance( [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] with excutils.save_and_reraise_exception(): [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] self.force_reraise() [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] raise self.value [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] updated_port = self._update_port( [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] _ensure_no_port_binding_failure(port) [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.212010] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] raise exception.PortBindingFailed(port_id=port['id']) [ 601.212290] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] nova.exception.PortBindingFailed: Binding failed for port 42cde311-8e8e-4b9a-a5eb-55168112473b, please check neutron logs for more information. [ 601.212290] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] [ 601.212290] env[61985]: INFO nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Terminating instance [ 601.216570] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Acquiring lock "refresh_cache-896ea04a-ad5b-40bb-b521-9788736012bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.216570] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Acquired lock "refresh_cache-896ea04a-ad5b-40bb-b521-9788736012bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.216570] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 601.319226] env[61985]: DEBUG nova.scheduler.client.report [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.342859] env[61985]: DEBUG nova.network.neutron [-] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.758935] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.826906] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.739s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.827983] env[61985]: ERROR nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e0c1bd46-c8f9-40bf-9baf-c9224e158573, please check neutron logs for more information. [ 601.827983] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] Traceback (most recent call last): [ 601.827983] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 601.827983] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] self.driver.spawn(context, instance, image_meta, [ 601.827983] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 601.827983] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.827983] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.827983] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] vm_ref = self.build_virtual_machine(instance, [ 601.827983] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.827983] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.827983] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] for vif in network_info: [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] return self._sync_wrapper(fn, *args, **kwargs) [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] self.wait() [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] self[:] = self._gt.wait() [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] return self._exit_event.wait() [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] result = hub.switch() [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.828592] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] return self.greenlet.switch() [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] result = function(*args, **kwargs) [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] return func(*args, **kwargs) [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] raise e [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] nwinfo = self.network_api.allocate_for_instance( [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] created_port_ids = self._update_ports_for_instance( [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] with excutils.save_and_reraise_exception(): [ 601.829314] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] self.force_reraise() [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] raise self.value [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] updated_port = self._update_port( [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] _ensure_no_port_binding_failure(port) [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] raise exception.PortBindingFailed(port_id=port['id']) [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] nova.exception.PortBindingFailed: Binding failed for port e0c1bd46-c8f9-40bf-9baf-c9224e158573, please check neutron logs for more information. [ 601.829842] env[61985]: ERROR nova.compute.manager [instance: 1cec2750-5d68-497c-8155-22afe8725809] [ 601.830314] env[61985]: DEBUG nova.compute.utils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Binding failed for port e0c1bd46-c8f9-40bf-9baf-c9224e158573, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 601.830314] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 1.302s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.832947] env[61985]: DEBUG nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Build of instance 1cec2750-5d68-497c-8155-22afe8725809 was re-scheduled: Binding failed for port e0c1bd46-c8f9-40bf-9baf-c9224e158573, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 601.833402] env[61985]: DEBUG nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 601.833641] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquiring lock "refresh_cache-1cec2750-5d68-497c-8155-22afe8725809" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.833783] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquired lock "refresh_cache-1cec2750-5d68-497c-8155-22afe8725809" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.833935] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 601.838568] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "fb7a840d-1aa8-425a-be50-ee31cf880a6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.838780] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "fb7a840d-1aa8-425a-be50-ee31cf880a6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.842605] env[61985]: INFO nova.compute.manager [-] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Took 1.05 seconds to deallocate network for instance. [ 601.844627] env[61985]: DEBUG nova.compute.claims [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 601.844795] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.952639] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.340985] env[61985]: DEBUG nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 602.369625] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.447770] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fcd1b89-fe0c-4482-af7b-8efe771d6499 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.455973] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Releasing lock "refresh_cache-896ea04a-ad5b-40bb-b521-9788736012bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.456394] env[61985]: DEBUG nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 602.456626] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 602.457488] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50bf74f5-a5df-4066-ab2b-94b348b9f1e9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.460142] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0481e00-84b0-4daa-9159-9fa9faf85482 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.493586] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77c32f0-da52-4d3e-aada-ba8766f404b7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.496710] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.500531] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067d9152-d6a8-4da2-b7fc-0529d0f12633 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.511655] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Releasing lock "refresh_cache-1cec2750-5d68-497c-8155-22afe8725809" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.511901] env[61985]: DEBUG nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 602.512104] env[61985]: DEBUG nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.512308] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 602.519587] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa1f774-4521-49af-97d8-134319cb0497 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.527289] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 896ea04a-ad5b-40bb-b521-9788736012bc could not be found. [ 602.527551] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 602.527785] env[61985]: INFO nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Took 0.07 seconds to destroy the instance on the hypervisor. [ 602.528060] env[61985]: DEBUG oslo.service.loopingcall [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.528589] env[61985]: DEBUG nova.compute.manager [-] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.528741] env[61985]: DEBUG nova.network.neutron [-] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 602.537964] env[61985]: DEBUG nova.compute.provider_tree [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.542375] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.563163] env[61985]: DEBUG nova.network.neutron [-] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.752490] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquiring lock "8eb535d6-b993-4453-ade0-ed6c36125690" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.752717] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lock "8eb535d6-b993-4453-ade0-ed6c36125690" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.947077] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.042322] env[61985]: DEBUG nova.scheduler.client.report [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.048455] env[61985]: DEBUG nova.network.neutron [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.064953] env[61985]: DEBUG nova.network.neutron [-] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.256125] env[61985]: DEBUG nova.compute.manager [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 603.307478] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquiring lock "5933751c-0aac-4575-aae9-f75904e7ea93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.307726] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "5933751c-0aac-4575-aae9-f75904e7ea93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.553704] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.721s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.553704] env[61985]: ERROR nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 43d12649-e1b7-42d1-a35e-35a76f24d507, please check neutron logs for more information. [ 603.553704] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Traceback (most recent call last): [ 603.553704] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 603.553704] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] self.driver.spawn(context, instance, image_meta, [ 603.553704] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 603.553704] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.553704] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.553704] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] vm_ref = self.build_virtual_machine(instance, [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] for vif in network_info: [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] return self._sync_wrapper(fn, *args, **kwargs) [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] self.wait() [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] self[:] = self._gt.wait() [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] return self._exit_event.wait() [ 603.554088] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] result = hub.switch() [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] return self.greenlet.switch() [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] result = function(*args, **kwargs) [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] return func(*args, **kwargs) [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] raise e [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] nwinfo = self.network_api.allocate_for_instance( [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.554403] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] created_port_ids = self._update_ports_for_instance( [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] with excutils.save_and_reraise_exception(): [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] self.force_reraise() [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] raise self.value [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] updated_port = self._update_port( [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] _ensure_no_port_binding_failure(port) [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.554726] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] raise exception.PortBindingFailed(port_id=port['id']) [ 603.555037] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] nova.exception.PortBindingFailed: Binding failed for port 43d12649-e1b7-42d1-a35e-35a76f24d507, please check neutron logs for more information. [ 603.555037] env[61985]: ERROR nova.compute.manager [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] [ 603.555037] env[61985]: DEBUG nova.compute.utils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Binding failed for port 43d12649-e1b7-42d1-a35e-35a76f24d507, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 603.556988] env[61985]: DEBUG nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Build of instance d5ea62be-7be5-4c9f-990c-5d0ccd768556 was re-scheduled: Binding failed for port 43d12649-e1b7-42d1-a35e-35a76f24d507, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 603.556988] env[61985]: DEBUG nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 603.556988] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Acquiring lock "refresh_cache-d5ea62be-7be5-4c9f-990c-5d0ccd768556" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.556988] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Acquired lock "refresh_cache-d5ea62be-7be5-4c9f-990c-5d0ccd768556" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.557296] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.559216] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 1.714s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.566485] env[61985]: INFO nova.compute.manager [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 1cec2750-5d68-497c-8155-22afe8725809] Took 1.05 seconds to deallocate network for instance. [ 603.567980] env[61985]: INFO nova.compute.manager [-] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Took 1.04 seconds to deallocate network for instance. [ 603.571703] env[61985]: DEBUG nova.compute.claims [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 603.575521] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.782085] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.809957] env[61985]: DEBUG nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 603.994955] env[61985]: DEBUG nova.compute.manager [req-35d7a585-e008-4672-9d69-d5d68c143b53 req-07f82035-f550-4c98-8a33-c3d1f258853a service nova] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Received event network-changed-42cde311-8e8e-4b9a-a5eb-55168112473b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 603.994955] env[61985]: DEBUG nova.compute.manager [req-35d7a585-e008-4672-9d69-d5d68c143b53 req-07f82035-f550-4c98-8a33-c3d1f258853a service nova] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Refreshing instance network info cache due to event network-changed-42cde311-8e8e-4b9a-a5eb-55168112473b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 603.994955] env[61985]: DEBUG oslo_concurrency.lockutils [req-35d7a585-e008-4672-9d69-d5d68c143b53 req-07f82035-f550-4c98-8a33-c3d1f258853a service nova] Acquiring lock "refresh_cache-896ea04a-ad5b-40bb-b521-9788736012bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.995385] env[61985]: DEBUG oslo_concurrency.lockutils [req-35d7a585-e008-4672-9d69-d5d68c143b53 req-07f82035-f550-4c98-8a33-c3d1f258853a service nova] Acquired lock "refresh_cache-896ea04a-ad5b-40bb-b521-9788736012bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.995385] env[61985]: DEBUG nova.network.neutron [req-35d7a585-e008-4672-9d69-d5d68c143b53 req-07f82035-f550-4c98-8a33-c3d1f258853a service nova] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Refreshing network info cache for port 42cde311-8e8e-4b9a-a5eb-55168112473b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 604.093250] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.188615] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a95ba4b-8d55-47a3-a878-f7c5898f8c1c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.196911] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fe4bb1-9358-45fb-b0a5-32424f6a1cf0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.228039] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d2fc0e-60f3-400e-b4df-cbcb5d7368dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.235544] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86207c3d-b01b-4657-93f6-18bcd652a49b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.249951] env[61985]: DEBUG nova.compute.provider_tree [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.255761] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.327897] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.537584] env[61985]: DEBUG nova.network.neutron [req-35d7a585-e008-4672-9d69-d5d68c143b53 req-07f82035-f550-4c98-8a33-c3d1f258853a service nova] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.603167] env[61985]: INFO nova.scheduler.client.report [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Deleted allocations for instance 1cec2750-5d68-497c-8155-22afe8725809 [ 604.761017] env[61985]: DEBUG nova.scheduler.client.report [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.762136] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Releasing lock "refresh_cache-d5ea62be-7be5-4c9f-990c-5d0ccd768556" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.763318] env[61985]: DEBUG nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 604.763868] env[61985]: DEBUG nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.764175] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.777297] env[61985]: DEBUG nova.network.neutron [req-35d7a585-e008-4672-9d69-d5d68c143b53 req-07f82035-f550-4c98-8a33-c3d1f258853a service nova] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.800144] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.115744] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ad4eb264-3054-4f78-bed0-712dd09bc057 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "1cec2750-5d68-497c-8155-22afe8725809" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.389s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.218818] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquiring lock "0777696f-109c-4fe1-a6d0-1498fd8e3639" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.220069] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "0777696f-109c-4fe1-a6d0-1498fd8e3639" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.267571] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.709s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.268299] env[61985]: ERROR nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1a9802c4-5560-461c-9375-3d5a0fc3e1b7, please check neutron logs for more information. [ 605.268299] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Traceback (most recent call last): [ 605.268299] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 605.268299] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] self.driver.spawn(context, instance, image_meta, [ 605.268299] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 605.268299] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.268299] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.268299] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] vm_ref = self.build_virtual_machine(instance, [ 605.268299] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.268299] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.268299] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] for vif in network_info: [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] return self._sync_wrapper(fn, *args, **kwargs) [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] self.wait() [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] self[:] = self._gt.wait() [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] return self._exit_event.wait() [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] result = hub.switch() [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.268735] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] return self.greenlet.switch() [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] result = function(*args, **kwargs) [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] return func(*args, **kwargs) [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] raise e [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] nwinfo = self.network_api.allocate_for_instance( [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] created_port_ids = self._update_ports_for_instance( [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] with excutils.save_and_reraise_exception(): [ 605.269093] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] self.force_reraise() [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] raise self.value [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] updated_port = self._update_port( [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] _ensure_no_port_binding_failure(port) [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] raise exception.PortBindingFailed(port_id=port['id']) [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] nova.exception.PortBindingFailed: Binding failed for port 1a9802c4-5560-461c-9375-3d5a0fc3e1b7, please check neutron logs for more information. [ 605.269435] env[61985]: ERROR nova.compute.manager [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] [ 605.269736] env[61985]: DEBUG nova.compute.utils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Binding failed for port 1a9802c4-5560-461c-9375-3d5a0fc3e1b7, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 605.270345] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.325s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.272071] env[61985]: INFO nova.compute.claims [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.280020] env[61985]: DEBUG nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Build of instance 1719bf7e-1f12-4f40-a56e-60d179790dce was re-scheduled: Binding failed for port 1a9802c4-5560-461c-9375-3d5a0fc3e1b7, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 605.280020] env[61985]: DEBUG nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 605.280020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Acquiring lock "refresh_cache-1719bf7e-1f12-4f40-a56e-60d179790dce" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.280020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Acquired lock "refresh_cache-1719bf7e-1f12-4f40-a56e-60d179790dce" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.280278] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.283919] env[61985]: DEBUG oslo_concurrency.lockutils [req-35d7a585-e008-4672-9d69-d5d68c143b53 req-07f82035-f550-4c98-8a33-c3d1f258853a service nova] Releasing lock "refresh_cache-896ea04a-ad5b-40bb-b521-9788736012bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.284253] env[61985]: DEBUG nova.compute.manager [req-35d7a585-e008-4672-9d69-d5d68c143b53 req-07f82035-f550-4c98-8a33-c3d1f258853a service nova] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Received event network-vif-deleted-42cde311-8e8e-4b9a-a5eb-55168112473b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 605.302829] env[61985]: DEBUG nova.network.neutron [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.721218] env[61985]: DEBUG nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.805689] env[61985]: INFO nova.compute.manager [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] [instance: d5ea62be-7be5-4c9f-990c-5d0ccd768556] Took 1.04 seconds to deallocate network for instance. [ 605.819434] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.017238] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.244852] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.450299] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23839c0-6088-4150-9cec-db0ca9613367 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.462871] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344190df-8eab-4588-939b-553eca6327a5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.513564] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Acquiring lock "0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.513792] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Lock "0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.519684] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce2c529-ea03-404d-9a78-12a01171a77b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.522169] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "a6dfe3a9-9745-4101-b55b-fc45c867ab27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.522397] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "a6dfe3a9-9745-4101-b55b-fc45c867ab27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.524432] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Releasing lock "refresh_cache-1719bf7e-1f12-4f40-a56e-60d179790dce" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.524702] env[61985]: DEBUG nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 606.524793] env[61985]: DEBUG nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.524968] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 606.532855] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b139c4-5571-42c7-81ff-ed07f396441c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.553922] env[61985]: DEBUG nova.compute.provider_tree [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.569295] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.872098] env[61985]: INFO nova.scheduler.client.report [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Deleted allocations for instance d5ea62be-7be5-4c9f-990c-5d0ccd768556 [ 607.016499] env[61985]: DEBUG nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 607.025734] env[61985]: DEBUG nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 607.057727] env[61985]: DEBUG nova.scheduler.client.report [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.072793] env[61985]: DEBUG nova.network.neutron [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.173899] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquiring lock "131c313b-1be0-4cac-8720-bd6ab3a0688a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.174152] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "131c313b-1be0-4cac-8720-bd6ab3a0688a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.391452] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a221a80a-2a55-47b7-895b-0a8ac5bf17e3 tempest-ServerExternalEventsTest-924724637 tempest-ServerExternalEventsTest-924724637-project-member] Lock "d5ea62be-7be5-4c9f-990c-5d0ccd768556" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.608s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.547687] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.551863] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.565718] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.295s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.566247] env[61985]: DEBUG nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 607.571768] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 3.996s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.576808] env[61985]: INFO nova.compute.manager [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] [instance: 1719bf7e-1f12-4f40-a56e-60d179790dce] Took 1.05 seconds to deallocate network for instance. [ 607.679603] env[61985]: DEBUG nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 607.738799] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Acquiring lock "e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.738986] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Lock "e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.077533] env[61985]: DEBUG nova.compute.utils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.081304] env[61985]: DEBUG nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 608.081499] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 608.202528] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.208408] env[61985]: DEBUG nova.policy [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e2994215f354ca5818a2a3aeac557dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bea886962434d1495fb140b375abde8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 608.241325] env[61985]: DEBUG nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.281120] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ced0b4-1a39-41b3-a139-c3cf3665f38c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.293517] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bd5f17-8a99-4d42-abc6-1fb1c0baea0e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.332841] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995cf361-6c90-4b41-aab0-fdee6a56b074 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.344995] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52d730f-3663-43e6-aaec-341d5ad11f00 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.362202] env[61985]: DEBUG nova.compute.provider_tree [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.582981] env[61985]: DEBUG nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 608.632555] env[61985]: INFO nova.scheduler.client.report [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Deleted allocations for instance 1719bf7e-1f12-4f40-a56e-60d179790dce [ 608.783642] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.829811] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquiring lock "5475fd04-5911-457d-85c6-30a41a8984cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.829941] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "5475fd04-5911-457d-85c6-30a41a8984cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.865871] env[61985]: DEBUG nova.scheduler.client.report [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.146285] env[61985]: DEBUG oslo_concurrency.lockutils [None req-15aa98ed-1de1-43be-8f57-dea83b3c6794 tempest-ServersAdminNegativeTestJSON-1153742242 tempest-ServersAdminNegativeTestJSON-1153742242-project-member] Lock "1719bf7e-1f12-4f40-a56e-60d179790dce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.981s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.306149] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Successfully created port: 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.371137] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.799s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.371783] env[61985]: ERROR nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 42cde311-8e8e-4b9a-a5eb-55168112473b, please check neutron logs for more information. [ 609.371783] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Traceback (most recent call last): [ 609.371783] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 609.371783] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] self.driver.spawn(context, instance, image_meta, [ 609.371783] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.371783] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.371783] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.371783] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] vm_ref = self.build_virtual_machine(instance, [ 609.371783] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.371783] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.371783] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] for vif in network_info: [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] return self._sync_wrapper(fn, *args, **kwargs) [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] self.wait() [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] self[:] = self._gt.wait() [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] return self._exit_event.wait() [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] result = hub.switch() [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.372385] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] return self.greenlet.switch() [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] result = function(*args, **kwargs) [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] return func(*args, **kwargs) [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] raise e [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] nwinfo = self.network_api.allocate_for_instance( [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] created_port_ids = self._update_ports_for_instance( [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] with excutils.save_and_reraise_exception(): [ 609.373026] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] self.force_reraise() [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] raise self.value [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] updated_port = self._update_port( [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] _ensure_no_port_binding_failure(port) [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] raise exception.PortBindingFailed(port_id=port['id']) [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] nova.exception.PortBindingFailed: Binding failed for port 42cde311-8e8e-4b9a-a5eb-55168112473b, please check neutron logs for more information. [ 609.373679] env[61985]: ERROR nova.compute.manager [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] [ 609.374131] env[61985]: DEBUG nova.compute.utils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Binding failed for port 42cde311-8e8e-4b9a-a5eb-55168112473b, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.376250] env[61985]: DEBUG nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Build of instance 896ea04a-ad5b-40bb-b521-9788736012bc was re-scheduled: Binding failed for port 42cde311-8e8e-4b9a-a5eb-55168112473b, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.377124] env[61985]: DEBUG nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 609.377548] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Acquiring lock "refresh_cache-896ea04a-ad5b-40bb-b521-9788736012bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.380024] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Acquired lock "refresh_cache-896ea04a-ad5b-40bb-b521-9788736012bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.380024] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.380024] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.597s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.382234] env[61985]: INFO nova.compute.claims [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.599516] env[61985]: DEBUG nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 609.625168] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.625389] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.625924] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.625924] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.625924] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.626080] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.626842] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.627213] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.627399] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.627595] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.627774] env[61985]: DEBUG nova.virt.hardware [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.628671] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3af543e-c957-4523-aeae-06d0a8fc6fe5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.638113] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856c3e92-3f4e-4275-af36-dee97318f6d8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.652495] env[61985]: DEBUG nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.946853] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.176260] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.330899] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.405190] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "81d8927f-16fe-4823-a52f-32d213da739d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.405712] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "81d8927f-16fe-4823-a52f-32d213da739d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.591444] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce70144a-a3eb-4ced-952c-db6acd6abde5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.600686] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adbc90af-cfdf-4a29-9cd6-9675fbeefb86 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.642859] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b8a0ef-5ad3-4580-91be-952b13042fc6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.651658] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95222474-0ef2-4e52-b98c-5dfc1944c3f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.667108] env[61985]: DEBUG nova.compute.provider_tree [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.835699] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Releasing lock "refresh_cache-896ea04a-ad5b-40bb-b521-9788736012bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.835962] env[61985]: DEBUG nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 610.836602] env[61985]: DEBUG nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.836602] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.068211] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.171170] env[61985]: DEBUG nova.scheduler.client.report [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.331298] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Acquiring lock "0947bda8-bee6-430b-bd79-eccd98d23cd8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.331516] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Lock "0947bda8-bee6-430b-bd79-eccd98d23cd8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.572510] env[61985]: DEBUG nova.network.neutron [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.678817] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.679484] env[61985]: DEBUG nova.compute.manager [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 611.682417] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.355s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.685078] env[61985]: INFO nova.compute.claims [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.077900] env[61985]: INFO nova.compute.manager [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] [instance: 896ea04a-ad5b-40bb-b521-9788736012bc] Took 1.24 seconds to deallocate network for instance. [ 612.193021] env[61985]: DEBUG nova.compute.utils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 612.193021] env[61985]: DEBUG nova.compute.manager [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Not allocating networking since 'none' was specified. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 612.340308] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Acquiring lock "9ac9f4b8-fcef-458a-913c-72cfc3322b6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.341350] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Lock "9ac9f4b8-fcef-458a-913c-72cfc3322b6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.698730] env[61985]: DEBUG nova.compute.manager [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 612.976578] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcdbf9c8-0302-4cc6-9624-dd571e83f8d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.993018] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c573630-5a92-4a77-a6ad-b14760a746b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.027989] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec40d585-fe6d-48a8-9735-6004a874e1af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.037546] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb64654e-ac19-4ae1-a785-07858952a6be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.053759] env[61985]: DEBUG nova.compute.provider_tree [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.132361] env[61985]: INFO nova.scheduler.client.report [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Deleted allocations for instance 896ea04a-ad5b-40bb-b521-9788736012bc [ 613.290240] env[61985]: ERROR nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9, please check neutron logs for more information. [ 613.290240] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 613.290240] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.290240] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 613.290240] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.290240] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 613.290240] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.290240] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 613.290240] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.290240] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 613.290240] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.290240] env[61985]: ERROR nova.compute.manager raise self.value [ 613.290240] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.290240] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 613.290240] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.290240] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 613.290773] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.290773] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 613.290773] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9, please check neutron logs for more information. [ 613.290773] env[61985]: ERROR nova.compute.manager [ 613.290773] env[61985]: Traceback (most recent call last): [ 613.290773] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 613.290773] env[61985]: listener.cb(fileno) [ 613.290773] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.290773] env[61985]: result = function(*args, **kwargs) [ 613.290773] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.290773] env[61985]: return func(*args, **kwargs) [ 613.290773] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.290773] env[61985]: raise e [ 613.290773] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.290773] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 613.290773] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.290773] env[61985]: created_port_ids = self._update_ports_for_instance( [ 613.290773] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.290773] env[61985]: with excutils.save_and_reraise_exception(): [ 613.290773] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.290773] env[61985]: self.force_reraise() [ 613.290773] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.290773] env[61985]: raise self.value [ 613.290773] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.290773] env[61985]: updated_port = self._update_port( [ 613.290773] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.290773] env[61985]: _ensure_no_port_binding_failure(port) [ 613.290773] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.290773] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 613.291628] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9, please check neutron logs for more information. [ 613.291628] env[61985]: Removing descriptor: 18 [ 613.291628] env[61985]: ERROR nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9, please check neutron logs for more information. [ 613.291628] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Traceback (most recent call last): [ 613.291628] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 613.291628] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] yield resources [ 613.291628] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 613.291628] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] self.driver.spawn(context, instance, image_meta, [ 613.291628] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 613.291628] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.291628] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.291628] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] vm_ref = self.build_virtual_machine(instance, [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] for vif in network_info: [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] return self._sync_wrapper(fn, *args, **kwargs) [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] self.wait() [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] self[:] = self._gt.wait() [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] return self._exit_event.wait() [ 613.291969] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] result = hub.switch() [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] return self.greenlet.switch() [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] result = function(*args, **kwargs) [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] return func(*args, **kwargs) [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] raise e [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] nwinfo = self.network_api.allocate_for_instance( [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.292401] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] created_port_ids = self._update_ports_for_instance( [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] with excutils.save_and_reraise_exception(): [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] self.force_reraise() [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] raise self.value [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] updated_port = self._update_port( [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] _ensure_no_port_binding_failure(port) [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.292743] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] raise exception.PortBindingFailed(port_id=port['id']) [ 613.293120] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] nova.exception.PortBindingFailed: Binding failed for port 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9, please check neutron logs for more information. [ 613.293120] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] [ 613.293120] env[61985]: INFO nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Terminating instance [ 613.297707] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "refresh_cache-fb7a840d-1aa8-425a-be50-ee31cf880a6d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.297757] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquired lock "refresh_cache-fb7a840d-1aa8-425a-be50-ee31cf880a6d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.297920] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 613.558319] env[61985]: DEBUG nova.scheduler.client.report [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.641272] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e50bba4-41a9-4caf-a95a-e2050b80e636 tempest-ImagesNegativeTestJSON-876693084 tempest-ImagesNegativeTestJSON-876693084-project-member] Lock "896ea04a-ad5b-40bb-b521-9788736012bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.820s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.712550] env[61985]: DEBUG nova.compute.manager [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 613.745331] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.745571] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.745720] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.745893] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.746580] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.746823] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.747105] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.748019] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.748019] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.748019] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.748019] env[61985]: DEBUG nova.virt.hardware [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.748788] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511e5df5-cc03-47b5-a61b-ee378bbd86b1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.756395] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquiring lock "7f32151e-7be0-40e2-b210-72765b7551bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.756676] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "7f32151e-7be0-40e2-b210-72765b7551bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.763198] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c467626-27b7-4ef4-b774-5e6101a790e0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.780048] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 613.789325] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 613.789647] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dcc95ec6-bbf2-49c7-a485-2297be3c5cd1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.805735] env[61985]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 613.806138] env[61985]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61985) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 613.806620] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 613.806972] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Creating folder: Project (e99c770cd2c342ff9f653133a846929a). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 613.809017] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-55222a8d-f8b4-4fb9-92f5-0c4718278be0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.819337] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Created folder: Project (e99c770cd2c342ff9f653133a846929a) in parent group-v211285. [ 613.820922] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Creating folder: Instances. Parent ref: group-v211290. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 613.820922] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c4f6b9c-5289-42a8-9c83-28b53634e9a0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.833314] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Created folder: Instances in parent group-v211290. [ 613.834454] env[61985]: DEBUG oslo.service.loopingcall [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.834565] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 613.834688] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7e33b39-8d5a-4c74-b5eb-31409e29cfbc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.858286] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.863344] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "535a5c48-c816-44f7-a58c-734878d90cc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.863611] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "535a5c48-c816-44f7-a58c-734878d90cc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.873102] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 613.873102] env[61985]: value = "task-935610" [ 613.873102] env[61985]: _type = "Task" [ 613.873102] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.880215] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935610, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.051497] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.064748] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.382s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.065311] env[61985]: DEBUG nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 614.072627] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.828s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.074126] env[61985]: INFO nova.compute.claims [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.146909] env[61985]: DEBUG nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.385835] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935610, 'name': CreateVM_Task, 'duration_secs': 0.362302} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.385955] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 614.387872] env[61985]: DEBUG oslo_vmware.service [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4722a1-b467-4409-a4b1-0e3129d16cd3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.395891] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.395891] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.396156] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 614.396479] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ade43ae-9f0c-4ca4-89ba-85c60d94609f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.402793] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 614.402793] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e15262-bb1e-a389-a793-f4d40df8f8d3" [ 614.402793] env[61985]: _type = "Task" [ 614.402793] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.417590] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e15262-bb1e-a389-a793-f4d40df8f8d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.554147] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Releasing lock "refresh_cache-fb7a840d-1aa8-425a-be50-ee31cf880a6d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.554546] env[61985]: DEBUG nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 614.554739] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 614.555042] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f7a105cb-5de5-48d1-ae5a-d09218dc4eb8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.569058] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b95a3d8-0e08-4f18-8ffc-157429292d91 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.585347] env[61985]: DEBUG nova.compute.utils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 614.588790] env[61985]: DEBUG nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 614.588967] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 614.607940] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fb7a840d-1aa8-425a-be50-ee31cf880a6d could not be found. [ 614.607940] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 614.607940] env[61985]: INFO nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 614.608220] env[61985]: DEBUG oslo.service.loopingcall [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.608424] env[61985]: DEBUG nova.compute.manager [-] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.608537] env[61985]: DEBUG nova.network.neutron [-] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 614.647102] env[61985]: DEBUG nova.network.neutron [-] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.673718] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.712636] env[61985]: DEBUG nova.policy [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27ebec01a2a64d25a15b4946041561c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bb8be0adb354f0682ad47d560a7c4b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 614.920347] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.920608] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 614.920840] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.920980] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.921467] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 614.921659] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d15e83a1-d175-4668-8453-1900b61b49ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.932904] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 614.933109] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 614.933900] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b27af3-e6c0-4d33-b925-0e6782fd6c5f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.946634] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec90501c-be26-46cf-a3f8-04ed3ae3a9b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.952945] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 614.952945] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c4f601-2814-c710-1602-341d7138f0e3" [ 614.952945] env[61985]: _type = "Task" [ 614.952945] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.963476] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c4f601-2814-c710-1602-341d7138f0e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.092718] env[61985]: DEBUG nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 615.142123] env[61985]: DEBUG nova.compute.manager [req-9244802e-33f9-4e98-8f02-de1fb1bce560 req-14700179-497f-48bf-9233-bae347694998 service nova] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Received event network-changed-2b817aeb-c77a-4b3c-8038-896b4b7b6bc9 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 615.142340] env[61985]: DEBUG nova.compute.manager [req-9244802e-33f9-4e98-8f02-de1fb1bce560 req-14700179-497f-48bf-9233-bae347694998 service nova] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Refreshing instance network info cache due to event network-changed-2b817aeb-c77a-4b3c-8038-896b4b7b6bc9. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 615.142548] env[61985]: DEBUG oslo_concurrency.lockutils [req-9244802e-33f9-4e98-8f02-de1fb1bce560 req-14700179-497f-48bf-9233-bae347694998 service nova] Acquiring lock "refresh_cache-fb7a840d-1aa8-425a-be50-ee31cf880a6d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.142679] env[61985]: DEBUG oslo_concurrency.lockutils [req-9244802e-33f9-4e98-8f02-de1fb1bce560 req-14700179-497f-48bf-9233-bae347694998 service nova] Acquired lock "refresh_cache-fb7a840d-1aa8-425a-be50-ee31cf880a6d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.142830] env[61985]: DEBUG nova.network.neutron [req-9244802e-33f9-4e98-8f02-de1fb1bce560 req-14700179-497f-48bf-9233-bae347694998 service nova] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Refreshing network info cache for port 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 615.150868] env[61985]: DEBUG nova.network.neutron [-] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.397478] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41dea9a-41f7-4eb6-900a-f2996de5b096 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.407625] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5589281-a4ae-4802-8896-952e2f65ab8a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.450160] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b6a45a-b992-416f-b4a5-5c79bb508cc6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.461065] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37cb2b10-1b9d-40cf-8ce8-f904b41adc98 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.467999] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Preparing fetch location {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 615.468318] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Creating directory with path [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 615.468591] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24ad4e91-5d7d-4035-a2c1-0f048087e7f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.478837] env[61985]: DEBUG nova.compute.provider_tree [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.497520] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Created directory with path [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 615.497520] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Fetch image to [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 615.497520] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Downloading image file data fe6d817b-6194-440f-988a-f4a94c580922 to [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk on the data store datastore2 {{(pid=61985) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 615.498106] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea379c71-31ae-476a-b1d8-4965e98bc0a4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.507500] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d3068a-20e7-4d0a-b7af-b76cab6a3256 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.518998] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c7b4eb-3112-4a96-a587-728ac8204e84 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.556572] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84df8b1-a0f7-4304-b832-f67db2d48fa5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.564173] env[61985]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-b5b46df7-8d38-4498-8c01-706274a2f619 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.658087] env[61985]: INFO nova.compute.manager [-] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Took 1.05 seconds to deallocate network for instance. [ 615.660425] env[61985]: DEBUG nova.compute.claims [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 615.660713] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.664932] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Downloading image file data fe6d817b-6194-440f-988a-f4a94c580922 to the data store datastore2 {{(pid=61985) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 615.718781] env[61985]: DEBUG nova.network.neutron [req-9244802e-33f9-4e98-8f02-de1fb1bce560 req-14700179-497f-48bf-9233-bae347694998 service nova] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.729597] env[61985]: DEBUG oslo_vmware.rw_handles [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61985) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 615.969680] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Successfully created port: ed853d00-29be-4555-b581-b75ace7a9105 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 615.983335] env[61985]: DEBUG nova.scheduler.client.report [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.085577] env[61985]: DEBUG nova.network.neutron [req-9244802e-33f9-4e98-8f02-de1fb1bce560 req-14700179-497f-48bf-9233-bae347694998 service nova] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.107615] env[61985]: DEBUG nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 616.122718] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.123169] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.150827] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.151084] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.151244] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.151422] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.151919] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.151919] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.151919] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.152119] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.152199] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.152352] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.152511] env[61985]: DEBUG nova.virt.hardware [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.153405] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06301f73-be9a-4bb8-8253-1d22c36f6761 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.164337] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8e08e2-8597-41a3-8dce-288b42edea73 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.488980] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.488980] env[61985]: DEBUG nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 616.493039] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.946s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.494977] env[61985]: INFO nova.compute.claims [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.510992] env[61985]: DEBUG oslo_vmware.rw_handles [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Completed reading data from the image iterator. {{(pid=61985) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 616.511233] env[61985]: DEBUG oslo_vmware.rw_handles [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 616.579718] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Downloaded image file data fe6d817b-6194-440f-988a-f4a94c580922 to vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk on the data store datastore2 {{(pid=61985) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 616.584825] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Caching image {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 616.584825] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Copying Virtual Disk [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk to [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 616.584825] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-794519c0-dc00-4ab7-8be3-d188d492ff06 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.589450] env[61985]: DEBUG oslo_concurrency.lockutils [req-9244802e-33f9-4e98-8f02-de1fb1bce560 req-14700179-497f-48bf-9233-bae347694998 service nova] Releasing lock "refresh_cache-fb7a840d-1aa8-425a-be50-ee31cf880a6d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.593054] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 616.593054] env[61985]: value = "task-935611" [ 616.593054] env[61985]: _type = "Task" [ 616.593054] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.602870] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935611, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.868247] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "955181b1-eb46-424c-8c79-055638dc9dac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.868247] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "955181b1-eb46-424c-8c79-055638dc9dac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.003311] env[61985]: DEBUG nova.compute.utils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.004173] env[61985]: DEBUG nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.004354] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 617.109513] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935611, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.143540] env[61985]: DEBUG nova.policy [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27ebec01a2a64d25a15b4946041561c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bb8be0adb354f0682ad47d560a7c4b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 617.516898] env[61985]: DEBUG nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 617.609392] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935611, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67036} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.609681] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Copied Virtual Disk [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk to [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 617.609895] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Deleting the datastore file [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 617.610208] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a7a9af2-045e-480e-a21c-bed9cd190d1a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.618127] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 617.618127] env[61985]: value = "task-935612" [ 617.618127] env[61985]: _type = "Task" [ 617.618127] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.738545] env[61985]: DEBUG nova.compute.manager [req-50025864-57cd-42e8-b129-bf6505e5dd21 req-dacde23e-5b74-4f4c-b04f-9f97cd506ce2 service nova] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Received event network-vif-deleted-2b817aeb-c77a-4b3c-8038-896b4b7b6bc9 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 617.823891] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Successfully created port: d31a402a-5ee9-45ed-95bc-0b287b4ac300 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.848608] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b705949-f08f-4814-890e-53fa229b9b2f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.859391] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2af6ced-8832-491b-81d9-b0135b88b350 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.898083] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4db33f-15fe-4c7d-97a8-8455ed7c0db5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.907876] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a3c52c5-92cc-450b-9227-0df3e7281089 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.925608] env[61985]: DEBUG nova.compute.provider_tree [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.130761] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935612, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086727} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.131045] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 618.131267] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Moving file from [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc/fe6d817b-6194-440f-988a-f4a94c580922 to [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922. {{(pid=61985) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 618.132285] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-3c0383db-aed5-4782-ae43-5e08e429453c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.140510] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 618.140510] env[61985]: value = "task-935613" [ 618.140510] env[61985]: _type = "Task" [ 618.140510] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.151857] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935613, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.432084] env[61985]: DEBUG nova.scheduler.client.report [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.528801] env[61985]: DEBUG nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 618.568336] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.568336] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.568336] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.568553] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.568553] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.568553] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.568935] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.569345] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.569662] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.569963] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.570274] env[61985]: DEBUG nova.virt.hardware [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.571429] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a5b1ec-5d58-4246-af31-0d06e9d4e177 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.580370] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe2386a-1f29-4173-81b4-432669b9b624 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.653250] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935613, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.033592} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.653720] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] File moved {{(pid=61985) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 618.656021] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Cleaning up location [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 618.656021] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Deleting the datastore file [datastore2] vmware_temp/4bc0b6ce-2202-40cb-9356-d630bd11bacc {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 618.656021] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f7c3dac-1dfd-49e0-b394-1df0c5b7f3c8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.667167] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 618.667167] env[61985]: value = "task-935614" [ 618.667167] env[61985]: _type = "Task" [ 618.667167] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.677074] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935614, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.940231] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.940231] env[61985]: DEBUG nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 618.941888] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.390s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.944425] env[61985]: INFO nova.compute.claims [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.178868] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935614, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028742} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.179392] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 619.180653] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e830ac4-138c-4289-8eea-d0a6e70ffd13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.188509] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 619.188509] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5268c3dc-6bd4-93de-0e8c-9728fbf9a022" [ 619.188509] env[61985]: _type = "Task" [ 619.188509] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.199819] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5268c3dc-6bd4-93de-0e8c-9728fbf9a022, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.454190] env[61985]: DEBUG nova.compute.utils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.456443] env[61985]: DEBUG nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 619.456741] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 619.503776] env[61985]: ERROR nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ed853d00-29be-4555-b581-b75ace7a9105, please check neutron logs for more information. [ 619.503776] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 619.503776] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.503776] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 619.503776] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.503776] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 619.503776] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.503776] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 619.503776] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.503776] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 619.503776] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.503776] env[61985]: ERROR nova.compute.manager raise self.value [ 619.503776] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.503776] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 619.503776] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.503776] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 619.504268] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.504268] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 619.504268] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ed853d00-29be-4555-b581-b75ace7a9105, please check neutron logs for more information. [ 619.504268] env[61985]: ERROR nova.compute.manager [ 619.504268] env[61985]: Traceback (most recent call last): [ 619.504268] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 619.504268] env[61985]: listener.cb(fileno) [ 619.504268] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.504268] env[61985]: result = function(*args, **kwargs) [ 619.504268] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.504268] env[61985]: return func(*args, **kwargs) [ 619.504268] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.504268] env[61985]: raise e [ 619.504268] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.504268] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 619.504268] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.504268] env[61985]: created_port_ids = self._update_ports_for_instance( [ 619.504268] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.504268] env[61985]: with excutils.save_and_reraise_exception(): [ 619.504268] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.504268] env[61985]: self.force_reraise() [ 619.504268] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.504268] env[61985]: raise self.value [ 619.504268] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.504268] env[61985]: updated_port = self._update_port( [ 619.504268] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.504268] env[61985]: _ensure_no_port_binding_failure(port) [ 619.504268] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.504268] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 619.505421] env[61985]: nova.exception.PortBindingFailed: Binding failed for port ed853d00-29be-4555-b581-b75ace7a9105, please check neutron logs for more information. [ 619.505421] env[61985]: Removing descriptor: 18 [ 619.505421] env[61985]: ERROR nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ed853d00-29be-4555-b581-b75ace7a9105, please check neutron logs for more information. [ 619.505421] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Traceback (most recent call last): [ 619.505421] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 619.505421] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] yield resources [ 619.505421] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 619.505421] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] self.driver.spawn(context, instance, image_meta, [ 619.505421] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 619.505421] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.505421] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.505421] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] vm_ref = self.build_virtual_machine(instance, [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] for vif in network_info: [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] return self._sync_wrapper(fn, *args, **kwargs) [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] self.wait() [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] self[:] = self._gt.wait() [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] return self._exit_event.wait() [ 619.505826] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] result = hub.switch() [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] return self.greenlet.switch() [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] result = function(*args, **kwargs) [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] return func(*args, **kwargs) [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] raise e [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] nwinfo = self.network_api.allocate_for_instance( [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.506212] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] created_port_ids = self._update_ports_for_instance( [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] with excutils.save_and_reraise_exception(): [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] self.force_reraise() [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] raise self.value [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] updated_port = self._update_port( [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] _ensure_no_port_binding_failure(port) [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.506567] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] raise exception.PortBindingFailed(port_id=port['id']) [ 619.506891] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] nova.exception.PortBindingFailed: Binding failed for port ed853d00-29be-4555-b581-b75ace7a9105, please check neutron logs for more information. [ 619.506891] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] [ 619.506891] env[61985]: INFO nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Terminating instance [ 619.508438] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquiring lock "refresh_cache-5933751c-0aac-4575-aae9-f75904e7ea93" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.508676] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquired lock "refresh_cache-5933751c-0aac-4575-aae9-f75904e7ea93" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.508890] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 619.702102] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5268c3dc-6bd4-93de-0e8c-9728fbf9a022, 'name': SearchDatastore_Task, 'duration_secs': 0.012088} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.702102] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.702102] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8eb535d6-b993-4453-ade0-ed6c36125690/8eb535d6-b993-4453-ade0-ed6c36125690.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 619.702102] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-766b4ee0-5b5c-4cf6-9b56-329e4838a6fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.710979] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 619.710979] env[61985]: value = "task-935615" [ 619.710979] env[61985]: _type = "Task" [ 619.710979] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.722065] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935615, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.731858] env[61985]: DEBUG nova.policy [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c766a8d1cdef4727ad43849905362cf0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '763682d56698490b86561fa276d58cc7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 619.961944] env[61985]: DEBUG nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 620.078851] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.224506] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935615, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.276060] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f6acc4-28de-4c6e-8473-ab9bc7532a9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.286729] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8c3977-10d2-4403-81d6-69266253ff53 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.322083] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1609b6f6-c7eb-4d21-ab35-39be87275689 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.330947] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce0c6ada-d328-46a4-8fba-3720fbe514ea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.345560] env[61985]: DEBUG nova.compute.provider_tree [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.497904] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Successfully created port: 92d4dd75-b8c2-4862-9873-742bf99bc765 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.543569] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "f86a5054-f30e-4868-9c50-2fe0d384b0b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.543845] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "f86a5054-f30e-4868-9c50-2fe0d384b0b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.662755] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.723604] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935615, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.520609} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.723908] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8eb535d6-b993-4453-ade0-ed6c36125690/8eb535d6-b993-4453-ade0-ed6c36125690.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 620.725277] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 620.725277] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a22755cd-774b-4f01-aa98-c17beb373905 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.734545] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 620.734545] env[61985]: value = "task-935616" [ 620.734545] env[61985]: _type = "Task" [ 620.734545] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.744555] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935616, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.849497] env[61985]: DEBUG nova.scheduler.client.report [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.915991] env[61985]: DEBUG nova.compute.manager [req-36789d1a-1bcc-4907-a7f7-c199ca4e0831 req-d846552c-77b8-434c-b3d4-40061fa3b064 service nova] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Received event network-changed-ed853d00-29be-4555-b581-b75ace7a9105 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 620.915991] env[61985]: DEBUG nova.compute.manager [req-36789d1a-1bcc-4907-a7f7-c199ca4e0831 req-d846552c-77b8-434c-b3d4-40061fa3b064 service nova] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Refreshing instance network info cache due to event network-changed-ed853d00-29be-4555-b581-b75ace7a9105. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 620.915991] env[61985]: DEBUG oslo_concurrency.lockutils [req-36789d1a-1bcc-4907-a7f7-c199ca4e0831 req-d846552c-77b8-434c-b3d4-40061fa3b064 service nova] Acquiring lock "refresh_cache-5933751c-0aac-4575-aae9-f75904e7ea93" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.978182] env[61985]: DEBUG nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 621.021783] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.021783] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.021783] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.022018] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.022018] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.022156] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.022367] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.022526] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.023257] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.023257] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.023257] env[61985]: DEBUG nova.virt.hardware [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.024132] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3db21c-7199-4dd6-a2bb-e237e5da9202 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.033841] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6906c530-4d7a-411d-8f09-25975054acd5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.166716] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Releasing lock "refresh_cache-5933751c-0aac-4575-aae9-f75904e7ea93" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.168047] env[61985]: DEBUG nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 621.168047] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 621.168375] env[61985]: DEBUG oslo_concurrency.lockutils [req-36789d1a-1bcc-4907-a7f7-c199ca4e0831 req-d846552c-77b8-434c-b3d4-40061fa3b064 service nova] Acquired lock "refresh_cache-5933751c-0aac-4575-aae9-f75904e7ea93" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.168686] env[61985]: DEBUG nova.network.neutron [req-36789d1a-1bcc-4907-a7f7-c199ca4e0831 req-d846552c-77b8-434c-b3d4-40061fa3b064 service nova] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Refreshing network info cache for port ed853d00-29be-4555-b581-b75ace7a9105 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 621.170153] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3833cbdb-a0c3-40f8-85b9-fc61dc82ebd5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.182160] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca24a00d-4e7d-404c-91fd-1d94d215fda4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.217506] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5933751c-0aac-4575-aae9-f75904e7ea93 could not be found. [ 621.217654] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 621.217842] env[61985]: INFO nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Took 0.05 seconds to destroy the instance on the hypervisor. [ 621.218687] env[61985]: DEBUG oslo.service.loopingcall [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.218687] env[61985]: DEBUG nova.compute.manager [-] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.218809] env[61985]: DEBUG nova.network.neutron [-] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 621.248658] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935616, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069333} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.248946] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 621.251932] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f3b22c-da32-4213-b6c8-e8b1e2be1e6a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.275816] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 8eb535d6-b993-4453-ade0-ed6c36125690/8eb535d6-b993-4453-ade0-ed6c36125690.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 621.275816] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdbcfc44-c404-4641-b35c-4be48421525e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.292490] env[61985]: DEBUG nova.network.neutron [-] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.303114] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 621.303114] env[61985]: value = "task-935617" [ 621.303114] env[61985]: _type = "Task" [ 621.303114] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.314960] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935617, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.356205] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.356205] env[61985]: DEBUG nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 621.358671] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.156s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.360091] env[61985]: INFO nova.compute.claims [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.747170] env[61985]: DEBUG nova.network.neutron [req-36789d1a-1bcc-4907-a7f7-c199ca4e0831 req-d846552c-77b8-434c-b3d4-40061fa3b064 service nova] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.795849] env[61985]: DEBUG nova.network.neutron [-] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.813997] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935617, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.873299] env[61985]: DEBUG nova.compute.utils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.878805] env[61985]: DEBUG nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 621.879034] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 621.996613] env[61985]: DEBUG nova.policy [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26d6f86b74ef4f5ca7af4c94176bdf35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e8e226d5c9240928ccc35013189235c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 622.299533] env[61985]: INFO nova.compute.manager [-] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Took 1.08 seconds to deallocate network for instance. [ 622.305712] env[61985]: DEBUG nova.compute.claims [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 622.306039] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.320766] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935617, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.388683] env[61985]: DEBUG nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 622.483041] env[61985]: DEBUG nova.network.neutron [req-36789d1a-1bcc-4907-a7f7-c199ca4e0831 req-d846552c-77b8-434c-b3d4-40061fa3b064 service nova] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.724254] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf1491f-990d-4146-ac69-6e132b076ea1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.732274] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eeaf376-be54-4e16-b4f5-3ed812c9bca0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.767355] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4d4c6d-3141-44af-a503-d130ad9c90e3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.782019] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6110736c-0bd8-4c64-8649-6c56051e0042 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.795965] env[61985]: DEBUG nova.compute.provider_tree [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.817606] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935617, 'name': ReconfigVM_Task, 'duration_secs': 1.284693} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.817606] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 8eb535d6-b993-4453-ade0-ed6c36125690/8eb535d6-b993-4453-ade0-ed6c36125690.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 622.818295] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-889b9f87-ca78-4fb4-b402-fa3d371866b2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.826845] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 622.826845] env[61985]: value = "task-935618" [ 622.826845] env[61985]: _type = "Task" [ 622.826845] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.837774] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935618, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.987283] env[61985]: DEBUG oslo_concurrency.lockutils [req-36789d1a-1bcc-4907-a7f7-c199ca4e0831 req-d846552c-77b8-434c-b3d4-40061fa3b064 service nova] Releasing lock "refresh_cache-5933751c-0aac-4575-aae9-f75904e7ea93" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.988367] env[61985]: DEBUG nova.compute.manager [req-36789d1a-1bcc-4907-a7f7-c199ca4e0831 req-d846552c-77b8-434c-b3d4-40061fa3b064 service nova] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Received event network-vif-deleted-ed853d00-29be-4555-b581-b75ace7a9105 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 623.248291] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Successfully created port: 4a291c97-27f0-4634-a2b4-a6e9dfb0d905 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.264263] env[61985]: ERROR nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d31a402a-5ee9-45ed-95bc-0b287b4ac300, please check neutron logs for more information. [ 623.264263] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 623.264263] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.264263] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 623.264263] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.264263] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 623.264263] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.264263] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 623.264263] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.264263] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 623.264263] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.264263] env[61985]: ERROR nova.compute.manager raise self.value [ 623.264263] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.264263] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 623.264263] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.264263] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 623.264718] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.264718] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 623.264718] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d31a402a-5ee9-45ed-95bc-0b287b4ac300, please check neutron logs for more information. [ 623.264718] env[61985]: ERROR nova.compute.manager [ 623.264718] env[61985]: Traceback (most recent call last): [ 623.264718] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 623.264718] env[61985]: listener.cb(fileno) [ 623.264718] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.264718] env[61985]: result = function(*args, **kwargs) [ 623.264718] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.264718] env[61985]: return func(*args, **kwargs) [ 623.264718] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.264718] env[61985]: raise e [ 623.264718] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.264718] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 623.264718] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.264718] env[61985]: created_port_ids = self._update_ports_for_instance( [ 623.264718] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.264718] env[61985]: with excutils.save_and_reraise_exception(): [ 623.264718] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.264718] env[61985]: self.force_reraise() [ 623.264718] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.264718] env[61985]: raise self.value [ 623.264718] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.264718] env[61985]: updated_port = self._update_port( [ 623.264718] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.264718] env[61985]: _ensure_no_port_binding_failure(port) [ 623.264718] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.264718] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 623.265455] env[61985]: nova.exception.PortBindingFailed: Binding failed for port d31a402a-5ee9-45ed-95bc-0b287b4ac300, please check neutron logs for more information. [ 623.265455] env[61985]: Removing descriptor: 17 [ 623.265455] env[61985]: ERROR nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d31a402a-5ee9-45ed-95bc-0b287b4ac300, please check neutron logs for more information. [ 623.265455] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Traceback (most recent call last): [ 623.265455] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 623.265455] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] yield resources [ 623.265455] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 623.265455] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] self.driver.spawn(context, instance, image_meta, [ 623.265455] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 623.265455] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.265455] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.265455] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] vm_ref = self.build_virtual_machine(instance, [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] for vif in network_info: [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] return self._sync_wrapper(fn, *args, **kwargs) [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] self.wait() [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] self[:] = self._gt.wait() [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] return self._exit_event.wait() [ 623.265813] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] result = hub.switch() [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] return self.greenlet.switch() [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] result = function(*args, **kwargs) [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] return func(*args, **kwargs) [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] raise e [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] nwinfo = self.network_api.allocate_for_instance( [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.266185] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] created_port_ids = self._update_ports_for_instance( [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] with excutils.save_and_reraise_exception(): [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] self.force_reraise() [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] raise self.value [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] updated_port = self._update_port( [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] _ensure_no_port_binding_failure(port) [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.266565] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] raise exception.PortBindingFailed(port_id=port['id']) [ 623.266967] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] nova.exception.PortBindingFailed: Binding failed for port d31a402a-5ee9-45ed-95bc-0b287b4ac300, please check neutron logs for more information. [ 623.266967] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] [ 623.266967] env[61985]: INFO nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Terminating instance [ 623.270856] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquiring lock "refresh_cache-0777696f-109c-4fe1-a6d0-1498fd8e3639" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.270856] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquired lock "refresh_cache-0777696f-109c-4fe1-a6d0-1498fd8e3639" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.270856] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.297538] env[61985]: DEBUG nova.scheduler.client.report [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.339625] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935618, 'name': Rename_Task, 'duration_secs': 0.157156} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.339910] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 623.340287] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4db11ddd-d77f-47e0-a272-58a380ab391a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.347669] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 623.347669] env[61985]: value = "task-935619" [ 623.347669] env[61985]: _type = "Task" [ 623.347669] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.360545] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935619, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.404948] env[61985]: DEBUG nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 623.441312] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.441574] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.441731] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.441919] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.442073] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.442271] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.443024] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.443024] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.443024] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.443024] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.443273] env[61985]: DEBUG nova.virt.hardware [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.444279] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c843151-679c-46cc-9d19-3e4ab7be317a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.455653] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e72b8b-ddb8-4c6b-85f0-a7f0e8872503 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.742635] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Acquiring lock "033766e8-92c0-498d-b3e8-37de5b46c20b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.742635] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Lock "033766e8-92c0-498d-b3e8-37de5b46c20b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.806334] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.806856] env[61985]: DEBUG nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 623.810311] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.027s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.812251] env[61985]: INFO nova.compute.claims [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 623.827093] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.863024] env[61985]: DEBUG oslo_vmware.api [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935619, 'name': PowerOnVM_Task, 'duration_secs': 0.483995} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.864335] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 623.864521] env[61985]: INFO nova.compute.manager [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Took 10.15 seconds to spawn the instance on the hypervisor. [ 623.864776] env[61985]: DEBUG nova.compute.manager [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 623.865818] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505964a6-2331-44b0-a992-d055db44751e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.904493] env[61985]: DEBUG nova.compute.manager [req-b1834f9c-66d6-43e5-945f-f7c745407adf req-d59c038e-3952-4229-a9da-fa1f9564913e service nova] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Received event network-changed-d31a402a-5ee9-45ed-95bc-0b287b4ac300 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 623.904992] env[61985]: DEBUG nova.compute.manager [req-b1834f9c-66d6-43e5-945f-f7c745407adf req-d59c038e-3952-4229-a9da-fa1f9564913e service nova] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Refreshing instance network info cache due to event network-changed-d31a402a-5ee9-45ed-95bc-0b287b4ac300. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 623.904992] env[61985]: DEBUG oslo_concurrency.lockutils [req-b1834f9c-66d6-43e5-945f-f7c745407adf req-d59c038e-3952-4229-a9da-fa1f9564913e service nova] Acquiring lock "refresh_cache-0777696f-109c-4fe1-a6d0-1498fd8e3639" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.320424] env[61985]: DEBUG nova.compute.utils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.324634] env[61985]: DEBUG nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 624.324634] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 624.371133] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.393910] env[61985]: INFO nova.compute.manager [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Took 20.63 seconds to build instance. [ 624.574889] env[61985]: DEBUG nova.policy [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02614aae1e6f421dab717cf4da5b6d25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ecfef1f8797e4c7e8ba35df351685e81', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.753334] env[61985]: ERROR nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 92d4dd75-b8c2-4862-9873-742bf99bc765, please check neutron logs for more information. [ 624.753334] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.753334] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.753334] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.753334] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.753334] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.753334] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.753334] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.753334] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.753334] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 624.753334] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.753334] env[61985]: ERROR nova.compute.manager raise self.value [ 624.753334] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.753334] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.753334] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.753334] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.754118] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.754118] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.754118] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 92d4dd75-b8c2-4862-9873-742bf99bc765, please check neutron logs for more information. [ 624.754118] env[61985]: ERROR nova.compute.manager [ 624.754118] env[61985]: Traceback (most recent call last): [ 624.754118] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.754118] env[61985]: listener.cb(fileno) [ 624.754118] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.754118] env[61985]: result = function(*args, **kwargs) [ 624.754118] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.754118] env[61985]: return func(*args, **kwargs) [ 624.754118] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.754118] env[61985]: raise e [ 624.754118] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.754118] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 624.754118] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.754118] env[61985]: created_port_ids = self._update_ports_for_instance( [ 624.754118] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.754118] env[61985]: with excutils.save_and_reraise_exception(): [ 624.754118] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.754118] env[61985]: self.force_reraise() [ 624.754118] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.754118] env[61985]: raise self.value [ 624.754118] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.754118] env[61985]: updated_port = self._update_port( [ 624.754118] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.754118] env[61985]: _ensure_no_port_binding_failure(port) [ 624.754118] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.754118] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.757827] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 92d4dd75-b8c2-4862-9873-742bf99bc765, please check neutron logs for more information. [ 624.757827] env[61985]: Removing descriptor: 14 [ 624.757827] env[61985]: ERROR nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 92d4dd75-b8c2-4862-9873-742bf99bc765, please check neutron logs for more information. [ 624.757827] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Traceback (most recent call last): [ 624.757827] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 624.757827] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] yield resources [ 624.757827] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 624.757827] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] self.driver.spawn(context, instance, image_meta, [ 624.757827] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 624.757827] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.757827] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.757827] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] vm_ref = self.build_virtual_machine(instance, [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] for vif in network_info: [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] return self._sync_wrapper(fn, *args, **kwargs) [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] self.wait() [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] self[:] = self._gt.wait() [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] return self._exit_event.wait() [ 624.758319] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] result = hub.switch() [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] return self.greenlet.switch() [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] result = function(*args, **kwargs) [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] return func(*args, **kwargs) [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] raise e [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] nwinfo = self.network_api.allocate_for_instance( [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.758670] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] created_port_ids = self._update_ports_for_instance( [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] with excutils.save_and_reraise_exception(): [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] self.force_reraise() [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] raise self.value [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] updated_port = self._update_port( [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] _ensure_no_port_binding_failure(port) [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.759023] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] raise exception.PortBindingFailed(port_id=port['id']) [ 624.759366] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] nova.exception.PortBindingFailed: Binding failed for port 92d4dd75-b8c2-4862-9873-742bf99bc765, please check neutron logs for more information. [ 624.759366] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] [ 624.759366] env[61985]: INFO nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Terminating instance [ 624.764708] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Acquiring lock "refresh_cache-0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.764708] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Acquired lock "refresh_cache-0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.764708] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 624.781846] env[61985]: DEBUG nova.compute.manager [req-ea8abd6f-f424-404e-a617-5c4b3108d39b req-0cdb6e4e-23d8-45dc-93f1-b46be23ea350 service nova] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Received event network-changed-92d4dd75-b8c2-4862-9873-742bf99bc765 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 624.783022] env[61985]: DEBUG nova.compute.manager [req-ea8abd6f-f424-404e-a617-5c4b3108d39b req-0cdb6e4e-23d8-45dc-93f1-b46be23ea350 service nova] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Refreshing instance network info cache due to event network-changed-92d4dd75-b8c2-4862-9873-742bf99bc765. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 624.783022] env[61985]: DEBUG oslo_concurrency.lockutils [req-ea8abd6f-f424-404e-a617-5c4b3108d39b req-0cdb6e4e-23d8-45dc-93f1-b46be23ea350 service nova] Acquiring lock "refresh_cache-0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.829467] env[61985]: DEBUG nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 624.873450] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Releasing lock "refresh_cache-0777696f-109c-4fe1-a6d0-1498fd8e3639" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.873854] env[61985]: DEBUG nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 624.880323] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 624.880323] env[61985]: DEBUG oslo_concurrency.lockutils [req-b1834f9c-66d6-43e5-945f-f7c745407adf req-d59c038e-3952-4229-a9da-fa1f9564913e service nova] Acquired lock "refresh_cache-0777696f-109c-4fe1-a6d0-1498fd8e3639" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.880323] env[61985]: DEBUG nova.network.neutron [req-b1834f9c-66d6-43e5-945f-f7c745407adf req-d59c038e-3952-4229-a9da-fa1f9564913e service nova] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Refreshing network info cache for port d31a402a-5ee9-45ed-95bc-0b287b4ac300 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 624.880323] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2113c0d-af4d-4ec3-bb72-041303896cbb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.896834] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0df420-b72b-4eba-9b1c-5366c82fb8b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.919080] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77ecc258-de0a-4e7d-a70d-0187b64e2dc9 tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lock "8eb535d6-b993-4453-ade0-ed6c36125690" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.166s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.936770] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0777696f-109c-4fe1-a6d0-1498fd8e3639 could not be found. [ 624.937143] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 624.937255] env[61985]: INFO nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Took 0.06 seconds to destroy the instance on the hypervisor. [ 624.937519] env[61985]: DEBUG oslo.service.loopingcall [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.937711] env[61985]: DEBUG nova.compute.manager [-] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.937799] env[61985]: DEBUG nova.network.neutron [-] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 624.983281] env[61985]: DEBUG nova.network.neutron [-] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.238055] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bff653-a391-4d2a-b63b-dc4e7408cdbd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.250184] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c804618-2396-4a48-92f3-55e67472f2e9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.298287] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39677fe9-698d-4358-af80-b87e71804146 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.308870] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6de34e4-465e-4813-9ba5-f2724efd67c8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.323894] env[61985]: DEBUG nova.compute.provider_tree [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.339560] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.422860] env[61985]: DEBUG nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.442191] env[61985]: DEBUG nova.network.neutron [req-b1834f9c-66d6-43e5-945f-f7c745407adf req-d59c038e-3952-4229-a9da-fa1f9564913e service nova] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.489076] env[61985]: DEBUG nova.network.neutron [-] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.541939] env[61985]: DEBUG nova.compute.manager [None req-81542c7e-f55a-446a-8fe6-1c6b40e80205 tempest-ServerDiagnosticsV248Test-779269505 tempest-ServerDiagnosticsV248Test-779269505-project-admin] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 625.543076] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15aad79d-9b40-4265-9426-c5307a028a65 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.555402] env[61985]: INFO nova.compute.manager [None req-81542c7e-f55a-446a-8fe6-1c6b40e80205 tempest-ServerDiagnosticsV248Test-779269505 tempest-ServerDiagnosticsV248Test-779269505-project-admin] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Retrieving diagnostics [ 625.556439] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606ff9ff-515e-4c0b-871d-f134a9006258 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.621382] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.829389] env[61985]: DEBUG nova.scheduler.client.report [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.851074] env[61985]: DEBUG nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 625.883832] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.884257] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.884427] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.884674] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.884843] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.885072] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.885374] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.885558] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.885791] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.886233] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.886696] env[61985]: DEBUG nova.virt.hardware [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.888563] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2f5624-2fb4-4a93-ba26-63176fc3423c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.900800] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5852d8c-8312-465d-a0c2-540aeee3909f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.949790] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.995751] env[61985]: INFO nova.compute.manager [-] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Took 1.06 seconds to deallocate network for instance. [ 625.999083] env[61985]: DEBUG nova.compute.claims [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 625.999269] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.014030] env[61985]: DEBUG nova.network.neutron [req-b1834f9c-66d6-43e5-945f-f7c745407adf req-d59c038e-3952-4229-a9da-fa1f9564913e service nova] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.123838] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Releasing lock "refresh_cache-0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.124419] env[61985]: DEBUG nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 626.124605] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 626.125584] env[61985]: DEBUG oslo_concurrency.lockutils [req-ea8abd6f-f424-404e-a617-5c4b3108d39b req-0cdb6e4e-23d8-45dc-93f1-b46be23ea350 service nova] Acquired lock "refresh_cache-0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.125584] env[61985]: DEBUG nova.network.neutron [req-ea8abd6f-f424-404e-a617-5c4b3108d39b req-0cdb6e4e-23d8-45dc-93f1-b46be23ea350 service nova] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Refreshing network info cache for port 92d4dd75-b8c2-4862-9873-742bf99bc765 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 626.126358] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-094ec84f-991c-41f1-b627-f3dcc41238f5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.140024] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9beb2ac-9c46-4da2-8a99-edc9632f0fb6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.171819] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba could not be found. [ 626.172273] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 626.172273] env[61985]: INFO nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Took 0.05 seconds to destroy the instance on the hypervisor. [ 626.172500] env[61985]: DEBUG oslo.service.loopingcall [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.172715] env[61985]: DEBUG nova.compute.manager [-] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.172810] env[61985]: DEBUG nova.network.neutron [-] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 626.334841] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.335418] env[61985]: DEBUG nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 626.340353] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.163s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.340353] env[61985]: INFO nova.compute.claims [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.344223] env[61985]: DEBUG nova.network.neutron [-] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.516512] env[61985]: DEBUG oslo_concurrency.lockutils [req-b1834f9c-66d6-43e5-945f-f7c745407adf req-d59c038e-3952-4229-a9da-fa1f9564913e service nova] Releasing lock "refresh_cache-0777696f-109c-4fe1-a6d0-1498fd8e3639" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.701476] env[61985]: ERROR nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4a291c97-27f0-4634-a2b4-a6e9dfb0d905, please check neutron logs for more information. [ 626.701476] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.701476] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.701476] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.701476] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.701476] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.701476] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.701476] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.701476] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.701476] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 626.701476] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.701476] env[61985]: ERROR nova.compute.manager raise self.value [ 626.701476] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.701476] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.701476] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.701476] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.702248] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.702248] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.702248] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4a291c97-27f0-4634-a2b4-a6e9dfb0d905, please check neutron logs for more information. [ 626.702248] env[61985]: ERROR nova.compute.manager [ 626.702248] env[61985]: Traceback (most recent call last): [ 626.702248] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.702248] env[61985]: listener.cb(fileno) [ 626.702248] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.702248] env[61985]: result = function(*args, **kwargs) [ 626.702248] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.702248] env[61985]: return func(*args, **kwargs) [ 626.702248] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.702248] env[61985]: raise e [ 626.702248] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.702248] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 626.702248] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.702248] env[61985]: created_port_ids = self._update_ports_for_instance( [ 626.702248] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.702248] env[61985]: with excutils.save_and_reraise_exception(): [ 626.702248] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.702248] env[61985]: self.force_reraise() [ 626.702248] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.702248] env[61985]: raise self.value [ 626.702248] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.702248] env[61985]: updated_port = self._update_port( [ 626.702248] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.702248] env[61985]: _ensure_no_port_binding_failure(port) [ 626.702248] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.702248] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.702992] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 4a291c97-27f0-4634-a2b4-a6e9dfb0d905, please check neutron logs for more information. [ 626.702992] env[61985]: Removing descriptor: 19 [ 626.702992] env[61985]: ERROR nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4a291c97-27f0-4634-a2b4-a6e9dfb0d905, please check neutron logs for more information. [ 626.702992] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Traceback (most recent call last): [ 626.702992] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 626.702992] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] yield resources [ 626.702992] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 626.702992] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] self.driver.spawn(context, instance, image_meta, [ 626.702992] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 626.702992] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.702992] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.702992] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] vm_ref = self.build_virtual_machine(instance, [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] for vif in network_info: [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] return self._sync_wrapper(fn, *args, **kwargs) [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] self.wait() [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] self[:] = self._gt.wait() [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] return self._exit_event.wait() [ 626.703323] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] result = hub.switch() [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] return self.greenlet.switch() [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] result = function(*args, **kwargs) [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] return func(*args, **kwargs) [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] raise e [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] nwinfo = self.network_api.allocate_for_instance( [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.703713] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] created_port_ids = self._update_ports_for_instance( [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] with excutils.save_and_reraise_exception(): [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] self.force_reraise() [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] raise self.value [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] updated_port = self._update_port( [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] _ensure_no_port_binding_failure(port) [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.704567] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] raise exception.PortBindingFailed(port_id=port['id']) [ 626.705143] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] nova.exception.PortBindingFailed: Binding failed for port 4a291c97-27f0-4634-a2b4-a6e9dfb0d905, please check neutron logs for more information. [ 626.705143] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] [ 626.705143] env[61985]: INFO nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Terminating instance [ 626.706331] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "refresh_cache-a6dfe3a9-9745-4101-b55b-fc45c867ab27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.706499] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquired lock "refresh_cache-a6dfe3a9-9745-4101-b55b-fc45c867ab27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.706668] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 626.712475] env[61985]: DEBUG nova.network.neutron [req-ea8abd6f-f424-404e-a617-5c4b3108d39b req-0cdb6e4e-23d8-45dc-93f1-b46be23ea350 service nova] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.716752] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Successfully created port: a40d81ad-693f-47da-9348-398e8d746538 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.842613] env[61985]: DEBUG nova.compute.utils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 626.843645] env[61985]: DEBUG nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 626.843933] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 626.850138] env[61985]: DEBUG nova.network.neutron [-] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.007766] env[61985]: DEBUG nova.policy [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bbf7459311e14262a054cef60df13c35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f2d760c818a44fbe8307c085417b2099', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 627.120873] env[61985]: DEBUG nova.network.neutron [req-ea8abd6f-f424-404e-a617-5c4b3108d39b req-0cdb6e4e-23d8-45dc-93f1-b46be23ea350 service nova] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.269958] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.351381] env[61985]: DEBUG nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 627.360248] env[61985]: INFO nova.compute.manager [-] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Took 1.19 seconds to deallocate network for instance. [ 627.363864] env[61985]: DEBUG nova.compute.claims [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.364302] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.433635] env[61985]: DEBUG nova.compute.manager [req-1067e086-34d8-4f00-9209-dcd0d45bec47 req-5ba159c4-da3a-4d4f-8f86-805f0e3bc5bf service nova] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Received event network-vif-deleted-d31a402a-5ee9-45ed-95bc-0b287b4ac300 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 627.624326] env[61985]: DEBUG oslo_concurrency.lockutils [req-ea8abd6f-f424-404e-a617-5c4b3108d39b req-0cdb6e4e-23d8-45dc-93f1-b46be23ea350 service nova] Releasing lock "refresh_cache-0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.727421] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd840461-c7d0-4542-9bec-3ec0243a297f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.735431] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8419d063-cb12-492e-8c7d-60e940ec1728 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.767741] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51568403-5fc3-46f0-80fe-6f8b0a283c9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.776752] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df34c2ca-d9c8-4789-b7e8-dd50e3a7a7e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.792024] env[61985]: DEBUG nova.compute.provider_tree [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.794158] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.056112] env[61985]: DEBUG nova.compute.manager [req-42d56512-b6ae-492d-8e43-eed2dfec9bde req-b50b0eae-8532-4ccd-be40-3b13375f0597 service nova] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Received event network-vif-deleted-92d4dd75-b8c2-4862-9873-742bf99bc765 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 628.056670] env[61985]: DEBUG nova.compute.manager [req-42d56512-b6ae-492d-8e43-eed2dfec9bde req-b50b0eae-8532-4ccd-be40-3b13375f0597 service nova] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Received event network-changed-4a291c97-27f0-4634-a2b4-a6e9dfb0d905 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 628.056670] env[61985]: DEBUG nova.compute.manager [req-42d56512-b6ae-492d-8e43-eed2dfec9bde req-b50b0eae-8532-4ccd-be40-3b13375f0597 service nova] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Refreshing instance network info cache due to event network-changed-4a291c97-27f0-4634-a2b4-a6e9dfb0d905. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 628.056769] env[61985]: DEBUG oslo_concurrency.lockutils [req-42d56512-b6ae-492d-8e43-eed2dfec9bde req-b50b0eae-8532-4ccd-be40-3b13375f0597 service nova] Acquiring lock "refresh_cache-a6dfe3a9-9745-4101-b55b-fc45c867ab27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.118943] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Successfully created port: 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.299117] env[61985]: DEBUG nova.scheduler.client.report [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.301820] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Releasing lock "refresh_cache-a6dfe3a9-9745-4101-b55b-fc45c867ab27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.302466] env[61985]: DEBUG nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 628.302771] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 628.303425] env[61985]: DEBUG oslo_concurrency.lockutils [req-42d56512-b6ae-492d-8e43-eed2dfec9bde req-b50b0eae-8532-4ccd-be40-3b13375f0597 service nova] Acquired lock "refresh_cache-a6dfe3a9-9745-4101-b55b-fc45c867ab27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.304425] env[61985]: DEBUG nova.network.neutron [req-42d56512-b6ae-492d-8e43-eed2dfec9bde req-b50b0eae-8532-4ccd-be40-3b13375f0597 service nova] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Refreshing network info cache for port 4a291c97-27f0-4634-a2b4-a6e9dfb0d905 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 628.306086] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb9ac81c-1895-45c8-83d6-617c22d6c8dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.322078] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf7a737-fa76-42b9-afac-248cab23b78d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.354144] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a6dfe3a9-9745-4101-b55b-fc45c867ab27 could not be found. [ 628.354715] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 628.355024] env[61985]: INFO nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Took 0.05 seconds to destroy the instance on the hypervisor. [ 628.355385] env[61985]: DEBUG oslo.service.loopingcall [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.355853] env[61985]: DEBUG nova.compute.manager [-] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.356079] env[61985]: DEBUG nova.network.neutron [-] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 628.374045] env[61985]: DEBUG nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 628.409369] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.412938] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.412938] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.412938] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.412938] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.412938] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.413307] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.413307] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.413307] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.413307] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.413307] env[61985]: DEBUG nova.virt.hardware [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.413516] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0224139-3cc7-4791-b3c3-2d8e5bdb98fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.419187] env[61985]: DEBUG nova.network.neutron [-] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.425351] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9608dd5f-55ca-4837-9bda-f56d0b919059 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.804601] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.805182] env[61985]: DEBUG nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 628.809278] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.136s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.811274] env[61985]: INFO nova.compute.claims [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.903101] env[61985]: DEBUG nova.network.neutron [req-42d56512-b6ae-492d-8e43-eed2dfec9bde req-b50b0eae-8532-4ccd-be40-3b13375f0597 service nova] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.924541] env[61985]: DEBUG nova.network.neutron [-] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.317431] env[61985]: DEBUG nova.compute.utils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.321488] env[61985]: DEBUG nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 629.321670] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 629.429402] env[61985]: INFO nova.compute.manager [-] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Took 1.07 seconds to deallocate network for instance. [ 629.433421] env[61985]: DEBUG nova.compute.claims [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 629.433421] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.485615] env[61985]: DEBUG nova.network.neutron [req-42d56512-b6ae-492d-8e43-eed2dfec9bde req-b50b0eae-8532-4ccd-be40-3b13375f0597 service nova] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.543206] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "213a3e19-5589-4261-96b0-69acfb6798ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.543418] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "213a3e19-5589-4261-96b0-69acfb6798ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.549692] env[61985]: DEBUG nova.policy [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02614aae1e6f421dab717cf4da5b6d25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ecfef1f8797e4c7e8ba35df351685e81', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 629.824686] env[61985]: DEBUG nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 629.988417] env[61985]: DEBUG oslo_concurrency.lockutils [req-42d56512-b6ae-492d-8e43-eed2dfec9bde req-b50b0eae-8532-4ccd-be40-3b13375f0597 service nova] Releasing lock "refresh_cache-a6dfe3a9-9745-4101-b55b-fc45c867ab27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.988732] env[61985]: DEBUG nova.compute.manager [req-42d56512-b6ae-492d-8e43-eed2dfec9bde req-b50b0eae-8532-4ccd-be40-3b13375f0597 service nova] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Received event network-vif-deleted-4a291c97-27f0-4634-a2b4-a6e9dfb0d905 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 630.167884] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba6353d-8dcf-44af-9097-f9548d9eaa53 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.178699] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc51a163-c409-4fcc-9579-10e051a3b2ea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.219818] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fafe195-d269-41ef-a17c-4fe82959f012 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.229918] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2930b2f7-8366-40ee-af53-c14198b928c9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.248790] env[61985]: DEBUG nova.compute.provider_tree [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.662310] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Acquiring lock "ba2fd184-e8cd-4667-8a41-eb1994c3329f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.662576] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Lock "ba2fd184-e8cd-4667-8a41-eb1994c3329f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.750987] env[61985]: DEBUG nova.scheduler.client.report [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.840150] env[61985]: DEBUG nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 630.879591] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 630.879591] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 630.879591] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 630.880280] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 630.880280] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 630.880280] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 630.880280] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 630.883417] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 630.883417] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 630.883417] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 630.883417] env[61985]: DEBUG nova.virt.hardware [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.884263] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42bf6b30-83fb-4711-97bf-d8944d0fdfd9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.894929] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9bf2c0-8d45-4f0f-90ec-679ea4b6a338 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.151015] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Successfully created port: ce0b4a7e-c317-4744-8740-3d762366dc34 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 631.258998] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.258998] env[61985]: DEBUG nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 631.261285] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.601s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.768921] env[61985]: DEBUG nova.compute.utils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 631.770117] env[61985]: DEBUG nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 631.770313] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 632.092016] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84168be5-f369-42f5-a14e-18bcba7234e2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.104035] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf60086-b56b-4274-a8ea-184d0e6957ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.135473] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb6abb7-126e-46e5-a6b3-8133a6fe26b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.145204] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee1c910-d0a0-45df-9e1e-6d87e2e73b58 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.163347] env[61985]: DEBUG nova.compute.provider_tree [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.246615] env[61985]: DEBUG nova.policy [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2dba42e753b4641862b3a03ce563892', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff085eb00ec4486b923bcdfd67406b14', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 632.270890] env[61985]: DEBUG nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 632.666197] env[61985]: DEBUG nova.scheduler.client.report [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.061226] env[61985]: ERROR nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b, please check neutron logs for more information. [ 633.061226] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 633.061226] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.061226] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 633.061226] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.061226] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 633.061226] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.061226] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 633.061226] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.061226] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 633.061226] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.061226] env[61985]: ERROR nova.compute.manager raise self.value [ 633.061226] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.061226] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 633.061226] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.061226] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 633.062181] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.062181] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 633.062181] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b, please check neutron logs for more information. [ 633.062181] env[61985]: ERROR nova.compute.manager [ 633.062181] env[61985]: Traceback (most recent call last): [ 633.062181] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 633.062181] env[61985]: listener.cb(fileno) [ 633.062181] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.062181] env[61985]: result = function(*args, **kwargs) [ 633.062181] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.062181] env[61985]: return func(*args, **kwargs) [ 633.062181] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.062181] env[61985]: raise e [ 633.062181] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.062181] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 633.062181] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.062181] env[61985]: created_port_ids = self._update_ports_for_instance( [ 633.062181] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.062181] env[61985]: with excutils.save_and_reraise_exception(): [ 633.062181] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.062181] env[61985]: self.force_reraise() [ 633.062181] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.062181] env[61985]: raise self.value [ 633.062181] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.062181] env[61985]: updated_port = self._update_port( [ 633.062181] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.062181] env[61985]: _ensure_no_port_binding_failure(port) [ 633.062181] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.062181] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 633.063199] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b, please check neutron logs for more information. [ 633.063199] env[61985]: Removing descriptor: 19 [ 633.063199] env[61985]: ERROR nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b, please check neutron logs for more information. [ 633.063199] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Traceback (most recent call last): [ 633.063199] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 633.063199] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] yield resources [ 633.063199] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 633.063199] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] self.driver.spawn(context, instance, image_meta, [ 633.063199] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 633.063199] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.063199] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.063199] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] vm_ref = self.build_virtual_machine(instance, [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] for vif in network_info: [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] return self._sync_wrapper(fn, *args, **kwargs) [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] self.wait() [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] self[:] = self._gt.wait() [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] return self._exit_event.wait() [ 633.063552] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] result = hub.switch() [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] return self.greenlet.switch() [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] result = function(*args, **kwargs) [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] return func(*args, **kwargs) [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] raise e [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] nwinfo = self.network_api.allocate_for_instance( [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.064238] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] created_port_ids = self._update_ports_for_instance( [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] with excutils.save_and_reraise_exception(): [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] self.force_reraise() [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] raise self.value [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] updated_port = self._update_port( [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] _ensure_no_port_binding_failure(port) [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.065572] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] raise exception.PortBindingFailed(port_id=port['id']) [ 633.066096] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] nova.exception.PortBindingFailed: Binding failed for port 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b, please check neutron logs for more information. [ 633.066096] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] [ 633.066096] env[61985]: INFO nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Terminating instance [ 633.067792] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Acquiring lock "refresh_cache-e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.068073] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Acquired lock "refresh_cache-e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.068187] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.173840] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.912s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.174811] env[61985]: ERROR nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9, please check neutron logs for more information. [ 633.174811] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Traceback (most recent call last): [ 633.174811] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 633.174811] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] self.driver.spawn(context, instance, image_meta, [ 633.174811] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 633.174811] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.174811] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.174811] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] vm_ref = self.build_virtual_machine(instance, [ 633.174811] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.174811] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.174811] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] for vif in network_info: [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] return self._sync_wrapper(fn, *args, **kwargs) [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] self.wait() [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] self[:] = self._gt.wait() [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] return self._exit_event.wait() [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] result = hub.switch() [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.175202] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] return self.greenlet.switch() [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] result = function(*args, **kwargs) [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] return func(*args, **kwargs) [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] raise e [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] nwinfo = self.network_api.allocate_for_instance( [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] created_port_ids = self._update_ports_for_instance( [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] with excutils.save_and_reraise_exception(): [ 633.175550] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] self.force_reraise() [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] raise self.value [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] updated_port = self._update_port( [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] _ensure_no_port_binding_failure(port) [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] raise exception.PortBindingFailed(port_id=port['id']) [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] nova.exception.PortBindingFailed: Binding failed for port 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9, please check neutron logs for more information. [ 633.175939] env[61985]: ERROR nova.compute.manager [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] [ 633.178160] env[61985]: DEBUG nova.compute.utils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Binding failed for port 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 633.178781] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.872s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.182327] env[61985]: DEBUG nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Build of instance fb7a840d-1aa8-425a-be50-ee31cf880a6d was re-scheduled: Binding failed for port 2b817aeb-c77a-4b3c-8038-896b4b7b6bc9, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 633.183681] env[61985]: DEBUG nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 633.184347] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "refresh_cache-fb7a840d-1aa8-425a-be50-ee31cf880a6d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.185234] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquired lock "refresh_cache-fb7a840d-1aa8-425a-be50-ee31cf880a6d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.188227] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.283568] env[61985]: DEBUG nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 633.327017] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.327017] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.327017] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.327377] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.327377] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.327377] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.327377] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.327377] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.327559] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.327559] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.328066] env[61985]: DEBUG nova.virt.hardware [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.329224] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2419ee02-2bc1-49d9-a312-7c0e0d619b20 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.340928] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb0f6e1-f941-429b-943e-4844dcce71be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.595699] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.648482] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "2e1e46ea-477c-41f3-938f-bd654bc5999d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.648482] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "2e1e46ea-477c-41f3-938f-bd654bc5999d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.677133] env[61985]: ERROR nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a40d81ad-693f-47da-9348-398e8d746538, please check neutron logs for more information. [ 633.677133] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 633.677133] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.677133] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 633.677133] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.677133] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 633.677133] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.677133] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 633.677133] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.677133] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 633.677133] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.677133] env[61985]: ERROR nova.compute.manager raise self.value [ 633.677133] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.677133] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 633.677133] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.677133] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 633.677646] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.677646] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 633.677646] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a40d81ad-693f-47da-9348-398e8d746538, please check neutron logs for more information. [ 633.677646] env[61985]: ERROR nova.compute.manager [ 633.677646] env[61985]: Traceback (most recent call last): [ 633.677646] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 633.677646] env[61985]: listener.cb(fileno) [ 633.677646] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.677646] env[61985]: result = function(*args, **kwargs) [ 633.677646] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.677646] env[61985]: return func(*args, **kwargs) [ 633.677646] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.677646] env[61985]: raise e [ 633.677646] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.677646] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 633.677646] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.677646] env[61985]: created_port_ids = self._update_ports_for_instance( [ 633.677646] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.677646] env[61985]: with excutils.save_and_reraise_exception(): [ 633.677646] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.677646] env[61985]: self.force_reraise() [ 633.677646] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.677646] env[61985]: raise self.value [ 633.677646] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.677646] env[61985]: updated_port = self._update_port( [ 633.677646] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.677646] env[61985]: _ensure_no_port_binding_failure(port) [ 633.677646] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.677646] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 633.678497] env[61985]: nova.exception.PortBindingFailed: Binding failed for port a40d81ad-693f-47da-9348-398e8d746538, please check neutron logs for more information. [ 633.678497] env[61985]: Removing descriptor: 18 [ 633.678497] env[61985]: ERROR nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a40d81ad-693f-47da-9348-398e8d746538, please check neutron logs for more information. [ 633.678497] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Traceback (most recent call last): [ 633.678497] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 633.678497] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] yield resources [ 633.678497] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 633.678497] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] self.driver.spawn(context, instance, image_meta, [ 633.678497] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 633.678497] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.678497] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.678497] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] vm_ref = self.build_virtual_machine(instance, [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] for vif in network_info: [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] return self._sync_wrapper(fn, *args, **kwargs) [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] self.wait() [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] self[:] = self._gt.wait() [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] return self._exit_event.wait() [ 633.678848] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] result = hub.switch() [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] return self.greenlet.switch() [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] result = function(*args, **kwargs) [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] return func(*args, **kwargs) [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] raise e [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] nwinfo = self.network_api.allocate_for_instance( [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.679348] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] created_port_ids = self._update_ports_for_instance( [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] with excutils.save_and_reraise_exception(): [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] self.force_reraise() [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] raise self.value [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] updated_port = self._update_port( [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] _ensure_no_port_binding_failure(port) [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.680953] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] raise exception.PortBindingFailed(port_id=port['id']) [ 633.681427] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] nova.exception.PortBindingFailed: Binding failed for port a40d81ad-693f-47da-9348-398e8d746538, please check neutron logs for more information. [ 633.681427] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] [ 633.681427] env[61985]: INFO nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Terminating instance [ 633.681427] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquiring lock "refresh_cache-131c313b-1be0-4cac-8720-bd6ab3a0688a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.681427] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquired lock "refresh_cache-131c313b-1be0-4cac-8720-bd6ab3a0688a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.681427] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.774599] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.017328] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.062118] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e550192-d8e9-4b35-86da-731c3d3d96c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.070397] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a513e6b5-f2d7-45ff-98d9-b4847b94d8a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.109773] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2d9899-701d-4f58-a9a2-c4d432fc62bc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.119128] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25427df-15f1-427d-921f-6d7bd5430b83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.137319] env[61985]: DEBUG nova.compute.provider_tree [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.232232] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.281222] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Releasing lock "refresh_cache-e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.281222] env[61985]: DEBUG nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 634.281222] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 634.282246] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a575f8b0-2589-45b6-978e-d8520372dc24 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.294937] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7872d1b-83f1-4a19-8bae-c68ad245ac15 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.315047] env[61985]: DEBUG nova.compute.manager [req-f27874f3-e2fc-4297-9851-3930cae7a9b8 req-9dc64059-64bd-4ea2-8a4f-abc24e04fd33 service nova] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Received event network-changed-567294bd-ed64-4faa-9dfd-b6c7afd9ac4b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 634.315576] env[61985]: DEBUG nova.compute.manager [req-f27874f3-e2fc-4297-9851-3930cae7a9b8 req-9dc64059-64bd-4ea2-8a4f-abc24e04fd33 service nova] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Refreshing instance network info cache due to event network-changed-567294bd-ed64-4faa-9dfd-b6c7afd9ac4b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 634.315576] env[61985]: DEBUG oslo_concurrency.lockutils [req-f27874f3-e2fc-4297-9851-3930cae7a9b8 req-9dc64059-64bd-4ea2-8a4f-abc24e04fd33 service nova] Acquiring lock "refresh_cache-e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.315725] env[61985]: DEBUG oslo_concurrency.lockutils [req-f27874f3-e2fc-4297-9851-3930cae7a9b8 req-9dc64059-64bd-4ea2-8a4f-abc24e04fd33 service nova] Acquired lock "refresh_cache-e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.315842] env[61985]: DEBUG nova.network.neutron [req-f27874f3-e2fc-4297-9851-3930cae7a9b8 req-9dc64059-64bd-4ea2-8a4f-abc24e04fd33 service nova] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Refreshing network info cache for port 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 634.323582] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Successfully created port: 02b62aa8-27c3-43b9-a22b-eb261089a8fb {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.333497] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c could not be found. [ 634.333790] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 634.334035] env[61985]: INFO nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 634.334334] env[61985]: DEBUG oslo.service.loopingcall [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.334616] env[61985]: DEBUG nova.compute.manager [-] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.334750] env[61985]: DEBUG nova.network.neutron [-] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.395890] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.424747] env[61985]: DEBUG nova.network.neutron [-] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.469798] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "40548796-fb8b-4b7e-9bfe-e83ca0edc01f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.470185] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "40548796-fb8b-4b7e-9bfe-e83ca0edc01f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.481072] env[61985]: ERROR nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ce0b4a7e-c317-4744-8740-3d762366dc34, please check neutron logs for more information. [ 634.481072] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.481072] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.481072] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.481072] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.481072] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.481072] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.481072] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.481072] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.481072] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 634.481072] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.481072] env[61985]: ERROR nova.compute.manager raise self.value [ 634.481072] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.481072] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.481072] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.481072] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.481554] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.481554] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.481554] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ce0b4a7e-c317-4744-8740-3d762366dc34, please check neutron logs for more information. [ 634.481554] env[61985]: ERROR nova.compute.manager [ 634.481554] env[61985]: Traceback (most recent call last): [ 634.481554] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.481554] env[61985]: listener.cb(fileno) [ 634.481554] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.481554] env[61985]: result = function(*args, **kwargs) [ 634.481554] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.481554] env[61985]: return func(*args, **kwargs) [ 634.481554] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.481554] env[61985]: raise e [ 634.481554] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.481554] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 634.481554] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.481554] env[61985]: created_port_ids = self._update_ports_for_instance( [ 634.481554] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.481554] env[61985]: with excutils.save_and_reraise_exception(): [ 634.481554] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.481554] env[61985]: self.force_reraise() [ 634.481554] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.481554] env[61985]: raise self.value [ 634.481554] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.481554] env[61985]: updated_port = self._update_port( [ 634.481554] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.481554] env[61985]: _ensure_no_port_binding_failure(port) [ 634.481554] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.481554] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.483420] env[61985]: nova.exception.PortBindingFailed: Binding failed for port ce0b4a7e-c317-4744-8740-3d762366dc34, please check neutron logs for more information. [ 634.483420] env[61985]: Removing descriptor: 14 [ 634.483420] env[61985]: ERROR nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ce0b4a7e-c317-4744-8740-3d762366dc34, please check neutron logs for more information. [ 634.483420] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Traceback (most recent call last): [ 634.483420] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 634.483420] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] yield resources [ 634.483420] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 634.483420] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] self.driver.spawn(context, instance, image_meta, [ 634.483420] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 634.483420] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.483420] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.483420] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] vm_ref = self.build_virtual_machine(instance, [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] for vif in network_info: [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] return self._sync_wrapper(fn, *args, **kwargs) [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] self.wait() [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] self[:] = self._gt.wait() [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] return self._exit_event.wait() [ 634.484259] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] result = hub.switch() [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] return self.greenlet.switch() [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] result = function(*args, **kwargs) [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] return func(*args, **kwargs) [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] raise e [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] nwinfo = self.network_api.allocate_for_instance( [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.484662] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] created_port_ids = self._update_ports_for_instance( [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] with excutils.save_and_reraise_exception(): [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] self.force_reraise() [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] raise self.value [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] updated_port = self._update_port( [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] _ensure_no_port_binding_failure(port) [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.485123] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] raise exception.PortBindingFailed(port_id=port['id']) [ 634.485794] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] nova.exception.PortBindingFailed: Binding failed for port ce0b4a7e-c317-4744-8740-3d762366dc34, please check neutron logs for more information. [ 634.485794] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] [ 634.485794] env[61985]: INFO nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Terminating instance [ 634.485794] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquiring lock "refresh_cache-5475fd04-5911-457d-85c6-30a41a8984cc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.485794] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquired lock "refresh_cache-5475fd04-5911-457d-85c6-30a41a8984cc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.485794] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 634.489396] env[61985]: DEBUG nova.compute.manager [req-969ea57c-c913-40a2-a2d1-f5e9d03703bd req-115b5bf1-91b0-495c-8dfa-28a15e423976 service nova] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Received event network-changed-a40d81ad-693f-47da-9348-398e8d746538 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 634.489554] env[61985]: DEBUG nova.compute.manager [req-969ea57c-c913-40a2-a2d1-f5e9d03703bd req-115b5bf1-91b0-495c-8dfa-28a15e423976 service nova] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Refreshing instance network info cache due to event network-changed-a40d81ad-693f-47da-9348-398e8d746538. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 634.489731] env[61985]: DEBUG oslo_concurrency.lockutils [req-969ea57c-c913-40a2-a2d1-f5e9d03703bd req-115b5bf1-91b0-495c-8dfa-28a15e423976 service nova] Acquiring lock "refresh_cache-131c313b-1be0-4cac-8720-bd6ab3a0688a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.527410] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.528040] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.552019] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.579607] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "983c68ad-52ca-4eba-9570-3a37d8f5c637" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.579691] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "983c68ad-52ca-4eba-9570-3a37d8f5c637" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.640725] env[61985]: DEBUG nova.scheduler.client.report [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.890215] env[61985]: DEBUG nova.network.neutron [req-f27874f3-e2fc-4297-9851-3930cae7a9b8 req-9dc64059-64bd-4ea2-8a4f-abc24e04fd33 service nova] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.898737] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Releasing lock "refresh_cache-131c313b-1be0-4cac-8720-bd6ab3a0688a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.899139] env[61985]: DEBUG nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 634.899322] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 634.901128] env[61985]: DEBUG oslo_concurrency.lockutils [req-969ea57c-c913-40a2-a2d1-f5e9d03703bd req-115b5bf1-91b0-495c-8dfa-28a15e423976 service nova] Acquired lock "refresh_cache-131c313b-1be0-4cac-8720-bd6ab3a0688a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.901258] env[61985]: DEBUG nova.network.neutron [req-969ea57c-c913-40a2-a2d1-f5e9d03703bd req-115b5bf1-91b0-495c-8dfa-28a15e423976 service nova] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Refreshing network info cache for port a40d81ad-693f-47da-9348-398e8d746538 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 634.902422] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b31ba22-7c24-4306-9e67-151d2e1f59c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.915149] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-accc18c4-71a5-416f-8504-0cc369c753aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.929121] env[61985]: DEBUG nova.network.neutron [-] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.948593] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 131c313b-1be0-4cac-8720-bd6ab3a0688a could not be found. [ 634.948776] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 634.948962] env[61985]: INFO nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 634.949232] env[61985]: DEBUG oslo.service.loopingcall [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.950321] env[61985]: DEBUG nova.compute.manager [-] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.950403] env[61985]: DEBUG nova.network.neutron [-] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.982789] env[61985]: DEBUG nova.network.neutron [-] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.006441] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.054951] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Releasing lock "refresh_cache-fb7a840d-1aa8-425a-be50-ee31cf880a6d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.055291] env[61985]: DEBUG nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 635.055391] env[61985]: DEBUG nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.055584] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 635.073294] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.127986] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.148021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.148021] env[61985]: ERROR nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ed853d00-29be-4555-b581-b75ace7a9105, please check neutron logs for more information. [ 635.148021] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Traceback (most recent call last): [ 635.148021] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 635.148021] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] self.driver.spawn(context, instance, image_meta, [ 635.148021] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.148021] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.148021] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.148021] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] vm_ref = self.build_virtual_machine(instance, [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] for vif in network_info: [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] return self._sync_wrapper(fn, *args, **kwargs) [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] self.wait() [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] self[:] = self._gt.wait() [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] return self._exit_event.wait() [ 635.148412] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] result = hub.switch() [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] return self.greenlet.switch() [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] result = function(*args, **kwargs) [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] return func(*args, **kwargs) [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] raise e [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] nwinfo = self.network_api.allocate_for_instance( [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.148832] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] created_port_ids = self._update_ports_for_instance( [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] with excutils.save_and_reraise_exception(): [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] self.force_reraise() [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] raise self.value [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] updated_port = self._update_port( [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] _ensure_no_port_binding_failure(port) [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.149238] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] raise exception.PortBindingFailed(port_id=port['id']) [ 635.149589] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] nova.exception.PortBindingFailed: Binding failed for port ed853d00-29be-4555-b581-b75ace7a9105, please check neutron logs for more information. [ 635.149589] env[61985]: ERROR nova.compute.manager [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] [ 635.149589] env[61985]: DEBUG nova.compute.utils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Binding failed for port ed853d00-29be-4555-b581-b75ace7a9105, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 635.152697] env[61985]: DEBUG nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Build of instance 5933751c-0aac-4575-aae9-f75904e7ea93 was re-scheduled: Binding failed for port ed853d00-29be-4555-b581-b75ace7a9105, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 635.152697] env[61985]: DEBUG nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 635.152697] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquiring lock "refresh_cache-5933751c-0aac-4575-aae9-f75904e7ea93" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.152697] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquired lock "refresh_cache-5933751c-0aac-4575-aae9-f75904e7ea93" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.153302] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.155150] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.206s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.159138] env[61985]: INFO nova.compute.claims [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.303886] env[61985]: DEBUG nova.network.neutron [req-f27874f3-e2fc-4297-9851-3930cae7a9b8 req-9dc64059-64bd-4ea2-8a4f-abc24e04fd33 service nova] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.436676] env[61985]: INFO nova.compute.manager [-] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Took 1.10 seconds to deallocate network for instance. [ 635.442544] env[61985]: DEBUG nova.compute.claims [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.443591] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.461307] env[61985]: DEBUG nova.network.neutron [req-969ea57c-c913-40a2-a2d1-f5e9d03703bd req-115b5bf1-91b0-495c-8dfa-28a15e423976 service nova] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.486237] env[61985]: DEBUG nova.network.neutron [-] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.576992] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Releasing lock "refresh_cache-5475fd04-5911-457d-85c6-30a41a8984cc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.577278] env[61985]: DEBUG nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 635.577715] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 635.578364] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3312befc-f71e-4a48-ba1c-0fc624d13e6c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.590378] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3835c48b-7567-4ef0-96e6-037bde854d93 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.627446] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5475fd04-5911-457d-85c6-30a41a8984cc could not be found. [ 635.627717] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 635.627904] env[61985]: INFO nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 635.628170] env[61985]: DEBUG oslo.service.loopingcall [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.628394] env[61985]: DEBUG nova.compute.manager [-] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.628489] env[61985]: DEBUG nova.network.neutron [-] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 635.631119] env[61985]: DEBUG nova.network.neutron [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.685080] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.802585] env[61985]: DEBUG nova.network.neutron [-] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.808474] env[61985]: DEBUG oslo_concurrency.lockutils [req-f27874f3-e2fc-4297-9851-3930cae7a9b8 req-9dc64059-64bd-4ea2-8a4f-abc24e04fd33 service nova] Releasing lock "refresh_cache-e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.826988] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.922162] env[61985]: DEBUG nova.network.neutron [req-969ea57c-c913-40a2-a2d1-f5e9d03703bd req-115b5bf1-91b0-495c-8dfa-28a15e423976 service nova] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.992351] env[61985]: INFO nova.compute.manager [-] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Took 1.04 seconds to deallocate network for instance. [ 635.998088] env[61985]: DEBUG nova.compute.claims [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.998297] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.093440] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "c8046008-3645-4845-be1e-23f5fefd5dc6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.093711] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "c8046008-3645-4845-be1e-23f5fefd5dc6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.134225] env[61985]: INFO nova.compute.manager [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: fb7a840d-1aa8-425a-be50-ee31cf880a6d] Took 1.08 seconds to deallocate network for instance. [ 636.307376] env[61985]: DEBUG nova.network.neutron [-] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.330959] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Releasing lock "refresh_cache-5933751c-0aac-4575-aae9-f75904e7ea93" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.331359] env[61985]: DEBUG nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 636.331549] env[61985]: DEBUG nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.331880] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.376741] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.424335] env[61985]: DEBUG oslo_concurrency.lockutils [req-969ea57c-c913-40a2-a2d1-f5e9d03703bd req-115b5bf1-91b0-495c-8dfa-28a15e423976 service nova] Releasing lock "refresh_cache-131c313b-1be0-4cac-8720-bd6ab3a0688a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.658078] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7112cf-68aa-476c-a15d-6c2d19e014d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.667472] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc71a903-6bd1-4d68-ac7e-1978aea7340b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.709293] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99e4a99-d1d2-4d25-a534-b6b779cb5172 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.723017] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ebc912c-36ee-48b7-bd9b-992f9106280e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.740956] env[61985]: DEBUG nova.compute.provider_tree [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.811427] env[61985]: INFO nova.compute.manager [-] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Took 1.18 seconds to deallocate network for instance. [ 636.815612] env[61985]: DEBUG nova.compute.claims [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 636.815666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.881757] env[61985]: DEBUG nova.network.neutron [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.188073] env[61985]: INFO nova.scheduler.client.report [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Deleted allocations for instance fb7a840d-1aa8-425a-be50-ee31cf880a6d [ 637.209016] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Acquiring lock "3480f0ec-85f4-4881-ad17-d647cd401fbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.210095] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Lock "3480f0ec-85f4-4881-ad17-d647cd401fbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.244410] env[61985]: DEBUG nova.scheduler.client.report [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.385442] env[61985]: INFO nova.compute.manager [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 5933751c-0aac-4575-aae9-f75904e7ea93] Took 1.05 seconds to deallocate network for instance. [ 637.685900] env[61985]: DEBUG nova.compute.manager [None req-43cc00e0-f43d-4461-9c7e-6a90fc7e6216 tempest-ServerDiagnosticsV248Test-779269505 tempest-ServerDiagnosticsV248Test-779269505-project-admin] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 637.689032] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8ac2c6-49f6-4d4b-af7a-8b2298c3f1c8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.703044] env[61985]: INFO nova.compute.manager [None req-43cc00e0-f43d-4461-9c7e-6a90fc7e6216 tempest-ServerDiagnosticsV248Test-779269505 tempest-ServerDiagnosticsV248Test-779269505-project-admin] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Retrieving diagnostics [ 637.705481] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acdc2b1a-7ff7-47ba-83f6-4fb63196f700 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.711225] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ff00adc-79bf-4ac8-942a-b49dec98804b tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "fb7a840d-1aa8-425a-be50-ee31cf880a6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.872s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.751441] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.596s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.751441] env[61985]: DEBUG nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 637.754033] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.755s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.850376] env[61985]: ERROR nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 02b62aa8-27c3-43b9-a22b-eb261089a8fb, please check neutron logs for more information. [ 637.850376] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.850376] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.850376] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.850376] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.850376] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.850376] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.850376] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.850376] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.850376] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 637.850376] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.850376] env[61985]: ERROR nova.compute.manager raise self.value [ 637.850376] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.850376] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.850376] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.850376] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.850862] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.850862] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.850862] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 02b62aa8-27c3-43b9-a22b-eb261089a8fb, please check neutron logs for more information. [ 637.850862] env[61985]: ERROR nova.compute.manager [ 637.850862] env[61985]: Traceback (most recent call last): [ 637.850862] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.850862] env[61985]: listener.cb(fileno) [ 637.850862] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.850862] env[61985]: result = function(*args, **kwargs) [ 637.850862] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.850862] env[61985]: return func(*args, **kwargs) [ 637.850862] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.850862] env[61985]: raise e [ 637.850862] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.850862] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 637.850862] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.850862] env[61985]: created_port_ids = self._update_ports_for_instance( [ 637.850862] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.850862] env[61985]: with excutils.save_and_reraise_exception(): [ 637.850862] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.850862] env[61985]: self.force_reraise() [ 637.850862] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.850862] env[61985]: raise self.value [ 637.850862] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.850862] env[61985]: updated_port = self._update_port( [ 637.850862] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.850862] env[61985]: _ensure_no_port_binding_failure(port) [ 637.850862] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.850862] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.851657] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 02b62aa8-27c3-43b9-a22b-eb261089a8fb, please check neutron logs for more information. [ 637.851657] env[61985]: Removing descriptor: 17 [ 637.851657] env[61985]: ERROR nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 02b62aa8-27c3-43b9-a22b-eb261089a8fb, please check neutron logs for more information. [ 637.851657] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Traceback (most recent call last): [ 637.851657] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 637.851657] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] yield resources [ 637.851657] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 637.851657] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] self.driver.spawn(context, instance, image_meta, [ 637.851657] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 637.851657] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.851657] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.851657] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] vm_ref = self.build_virtual_machine(instance, [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] for vif in network_info: [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] return self._sync_wrapper(fn, *args, **kwargs) [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] self.wait() [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] self[:] = self._gt.wait() [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] return self._exit_event.wait() [ 637.852112] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] result = hub.switch() [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] return self.greenlet.switch() [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] result = function(*args, **kwargs) [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] return func(*args, **kwargs) [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] raise e [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] nwinfo = self.network_api.allocate_for_instance( [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.852604] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] created_port_ids = self._update_ports_for_instance( [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] with excutils.save_and_reraise_exception(): [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] self.force_reraise() [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] raise self.value [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] updated_port = self._update_port( [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] _ensure_no_port_binding_failure(port) [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.852949] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] raise exception.PortBindingFailed(port_id=port['id']) [ 637.853320] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] nova.exception.PortBindingFailed: Binding failed for port 02b62aa8-27c3-43b9-a22b-eb261089a8fb, please check neutron logs for more information. [ 637.853320] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] [ 637.853320] env[61985]: INFO nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Terminating instance [ 637.854312] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "refresh_cache-81d8927f-16fe-4823-a52f-32d213da739d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.854312] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquired lock "refresh_cache-81d8927f-16fe-4823-a52f-32d213da739d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.854521] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 637.872142] env[61985]: DEBUG nova.compute.manager [req-878513a1-ad67-4d16-8a1d-11c8eb807fea req-55e975af-fc16-40af-885a-cb0b269990d0 service nova] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Received event network-vif-deleted-567294bd-ed64-4faa-9dfd-b6c7afd9ac4b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 637.872142] env[61985]: DEBUG nova.compute.manager [req-878513a1-ad67-4d16-8a1d-11c8eb807fea req-55e975af-fc16-40af-885a-cb0b269990d0 service nova] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Received event network-changed-ce0b4a7e-c317-4744-8740-3d762366dc34 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 637.872142] env[61985]: DEBUG nova.compute.manager [req-878513a1-ad67-4d16-8a1d-11c8eb807fea req-55e975af-fc16-40af-885a-cb0b269990d0 service nova] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Refreshing instance network info cache due to event network-changed-ce0b4a7e-c317-4744-8740-3d762366dc34. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 637.872142] env[61985]: DEBUG oslo_concurrency.lockutils [req-878513a1-ad67-4d16-8a1d-11c8eb807fea req-55e975af-fc16-40af-885a-cb0b269990d0 service nova] Acquiring lock "refresh_cache-5475fd04-5911-457d-85c6-30a41a8984cc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.872142] env[61985]: DEBUG oslo_concurrency.lockutils [req-878513a1-ad67-4d16-8a1d-11c8eb807fea req-55e975af-fc16-40af-885a-cb0b269990d0 service nova] Acquired lock "refresh_cache-5475fd04-5911-457d-85c6-30a41a8984cc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.872485] env[61985]: DEBUG nova.network.neutron [req-878513a1-ad67-4d16-8a1d-11c8eb807fea req-55e975af-fc16-40af-885a-cb0b269990d0 service nova] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Refreshing network info cache for port ce0b4a7e-c317-4744-8740-3d762366dc34 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 637.906685] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Acquiring lock "4a5bd6dd-9d21-43a2-9108-58aed0637ea9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.906911] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Lock "4a5bd6dd-9d21-43a2-9108-58aed0637ea9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.907129] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Acquiring lock "9af10e29-12ea-441d-b82a-be6976032c88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.907314] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Lock "9af10e29-12ea-441d-b82a-be6976032c88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.023594] env[61985]: DEBUG nova.compute.manager [req-ffffb51f-e588-443d-98ad-fd3958046b78 req-c28ba97f-a013-4d7c-a6f5-fe501244f419 service nova] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Received event network-vif-deleted-a40d81ad-693f-47da-9348-398e8d746538 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 638.216733] env[61985]: DEBUG nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 638.270446] env[61985]: DEBUG nova.compute.utils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 638.273525] env[61985]: DEBUG nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 638.273525] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 638.383746] env[61985]: DEBUG nova.policy [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '353319cf369a418787a8bade5e467cc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '18b736c242b247299c0cfa9ef2e96891', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 638.399019] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.413907] env[61985]: DEBUG nova.network.neutron [req-878513a1-ad67-4d16-8a1d-11c8eb807fea req-55e975af-fc16-40af-885a-cb0b269990d0 service nova] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.423190] env[61985]: INFO nova.scheduler.client.report [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Deleted allocations for instance 5933751c-0aac-4575-aae9-f75904e7ea93 [ 638.532463] env[61985]: DEBUG nova.network.neutron [req-878513a1-ad67-4d16-8a1d-11c8eb807fea req-55e975af-fc16-40af-885a-cb0b269990d0 service nova] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.683031] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.735307] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea0e15a-3032-4d7f-9122-2272e8e0d0d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.742979] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.751779] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7224007f-0705-4f02-a28f-52082dd2bdc2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.790470] env[61985]: DEBUG nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 638.794250] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e14941-e9c1-4618-a749-a9f3938787e2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.804086] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233a4310-c581-4d27-8e4a-427cf093b01a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.823246] env[61985]: DEBUG nova.compute.provider_tree [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.936921] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ef7c46d3-42d8-4046-8edf-12bef6552716 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "5933751c-0aac-4575-aae9-f75904e7ea93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.629s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.034933] env[61985]: DEBUG oslo_concurrency.lockutils [req-878513a1-ad67-4d16-8a1d-11c8eb807fea req-55e975af-fc16-40af-885a-cb0b269990d0 service nova] Releasing lock "refresh_cache-5475fd04-5911-457d-85c6-30a41a8984cc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.035244] env[61985]: DEBUG nova.compute.manager [req-878513a1-ad67-4d16-8a1d-11c8eb807fea req-55e975af-fc16-40af-885a-cb0b269990d0 service nova] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Received event network-vif-deleted-ce0b4a7e-c317-4744-8740-3d762366dc34 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 639.188757] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Releasing lock "refresh_cache-81d8927f-16fe-4823-a52f-32d213da739d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.188757] env[61985]: DEBUG nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 639.188757] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 639.188757] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c94fc9da-d256-44ce-b8c0-3fe26cfe111e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.200288] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e22fc58-4571-409a-9f93-2d4717e9f786 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.212242] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Successfully created port: 6b958751-3b84-41f9-b82b-5e37e471381a {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 639.231534] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 81d8927f-16fe-4823-a52f-32d213da739d could not be found. [ 639.231823] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 639.232026] env[61985]: INFO nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 639.232275] env[61985]: DEBUG oslo.service.loopingcall [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.232496] env[61985]: DEBUG nova.compute.manager [-] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.232579] env[61985]: DEBUG nova.network.neutron [-] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 639.264607] env[61985]: DEBUG nova.network.neutron [-] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.299608] env[61985]: INFO nova.virt.block_device [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Booting with volume 27d3a498-e600-4081-8c6c-2c22360d4922 at /dev/sda [ 639.325912] env[61985]: DEBUG nova.scheduler.client.report [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.360181] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-583fa886-aa0a-4186-aa56-4fc59eb43b0d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.370620] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2927cfaa-beff-4f6e-826c-0731486fa6b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.399377] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b39abc6f-4f6d-48e1-9dae-82ef77d053d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.407690] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7d6afe-69d1-4c0b-be86-f3d5199770ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.435782] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f62f6e-8de5-4121-a746-35b1fb67cc8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.440465] env[61985]: DEBUG nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 639.446526] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7861f7c6-b1ef-4dca-9934-d09ccb5e4ddd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.466100] env[61985]: DEBUG nova.virt.block_device [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Updating existing volume attachment record: 2ecb74d3-65e7-41bd-b8a7-efa60335917b {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 639.731881] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Acquiring lock "a0fd407f-7cf0-4c01-8195-399d8cbf62f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.732332] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Lock "a0fd407f-7cf0-4c01-8195-399d8cbf62f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.767490] env[61985]: DEBUG nova.network.neutron [-] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.835038] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.081s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.837897] env[61985]: ERROR nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d31a402a-5ee9-45ed-95bc-0b287b4ac300, please check neutron logs for more information. [ 639.837897] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Traceback (most recent call last): [ 639.837897] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 639.837897] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] self.driver.spawn(context, instance, image_meta, [ 639.837897] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 639.837897] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.837897] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.837897] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] vm_ref = self.build_virtual_machine(instance, [ 639.837897] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.837897] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.837897] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] for vif in network_info: [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] return self._sync_wrapper(fn, *args, **kwargs) [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] self.wait() [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] self[:] = self._gt.wait() [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] return self._exit_event.wait() [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] result = hub.switch() [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 639.838546] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] return self.greenlet.switch() [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] result = function(*args, **kwargs) [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] return func(*args, **kwargs) [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] raise e [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] nwinfo = self.network_api.allocate_for_instance( [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] created_port_ids = self._update_ports_for_instance( [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] with excutils.save_and_reraise_exception(): [ 639.838964] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] self.force_reraise() [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] raise self.value [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] updated_port = self._update_port( [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] _ensure_no_port_binding_failure(port) [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] raise exception.PortBindingFailed(port_id=port['id']) [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] nova.exception.PortBindingFailed: Binding failed for port d31a402a-5ee9-45ed-95bc-0b287b4ac300, please check neutron logs for more information. [ 639.839377] env[61985]: ERROR nova.compute.manager [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] [ 639.839838] env[61985]: DEBUG nova.compute.utils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Binding failed for port d31a402a-5ee9-45ed-95bc-0b287b4ac300, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 639.839838] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.475s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.843869] env[61985]: DEBUG nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Build of instance 0777696f-109c-4fe1-a6d0-1498fd8e3639 was re-scheduled: Binding failed for port d31a402a-5ee9-45ed-95bc-0b287b4ac300, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 639.844446] env[61985]: DEBUG nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 639.844622] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquiring lock "refresh_cache-0777696f-109c-4fe1-a6d0-1498fd8e3639" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.844771] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Acquired lock "refresh_cache-0777696f-109c-4fe1-a6d0-1498fd8e3639" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.844927] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.853878] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Acquiring lock "50315640-7609-4fff-b191-aa29f06cb4f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.856579] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Lock "50315640-7609-4fff-b191-aa29f06cb4f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.971677] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.191509] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquiring lock "8eb535d6-b993-4453-ade0-ed6c36125690" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.191509] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lock "8eb535d6-b993-4453-ade0-ed6c36125690" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.191509] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquiring lock "8eb535d6-b993-4453-ade0-ed6c36125690-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.191509] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lock "8eb535d6-b993-4453-ade0-ed6c36125690-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.191985] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lock "8eb535d6-b993-4453-ade0-ed6c36125690-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.194150] env[61985]: INFO nova.compute.manager [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Terminating instance [ 640.195884] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquiring lock "refresh_cache-8eb535d6-b993-4453-ade0-ed6c36125690" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.196043] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquired lock "refresh_cache-8eb535d6-b993-4453-ade0-ed6c36125690" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.196241] env[61985]: DEBUG nova.network.neutron [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 640.271716] env[61985]: INFO nova.compute.manager [-] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Took 1.04 seconds to deallocate network for instance. [ 640.276019] env[61985]: DEBUG nova.compute.claims [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.276019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.396265] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.601749] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.738804] env[61985]: DEBUG nova.network.neutron [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.794214] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9c2517-48c6-4c78-802a-13254b30b3ad {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.798719] env[61985]: ERROR nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6b958751-3b84-41f9-b82b-5e37e471381a, please check neutron logs for more information. [ 640.798719] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.798719] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.798719] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.798719] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.798719] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.798719] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.798719] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.798719] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.798719] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 640.798719] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.798719] env[61985]: ERROR nova.compute.manager raise self.value [ 640.798719] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.798719] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.798719] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.798719] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.799237] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.799237] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.799237] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6b958751-3b84-41f9-b82b-5e37e471381a, please check neutron logs for more information. [ 640.799237] env[61985]: ERROR nova.compute.manager [ 640.799711] env[61985]: Traceback (most recent call last): [ 640.799853] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.799853] env[61985]: listener.cb(fileno) [ 640.799954] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.799954] env[61985]: result = function(*args, **kwargs) [ 640.800052] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.800052] env[61985]: return func(*args, **kwargs) [ 640.800127] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.800127] env[61985]: raise e [ 640.800210] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.800210] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 640.800279] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.800279] env[61985]: created_port_ids = self._update_ports_for_instance( [ 640.800348] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.800348] env[61985]: with excutils.save_and_reraise_exception(): [ 640.800414] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.800414] env[61985]: self.force_reraise() [ 640.800479] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.800479] env[61985]: raise self.value [ 640.800547] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.800547] env[61985]: updated_port = self._update_port( [ 640.800615] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.800615] env[61985]: _ensure_no_port_binding_failure(port) [ 640.800690] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.800690] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.800769] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 6b958751-3b84-41f9-b82b-5e37e471381a, please check neutron logs for more information. [ 640.800818] env[61985]: Removing descriptor: 17 [ 640.807102] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfde9c3-321a-4a02-95fb-134140d7bd43 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.845018] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e586c2-6a60-4b37-983e-836ee403dc3c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.860187] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3fc485-cd7b-4f16-a78a-27e917177028 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.866339] env[61985]: DEBUG nova.network.neutron [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.878357] env[61985]: DEBUG nova.compute.provider_tree [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.942870] env[61985]: DEBUG nova.compute.manager [req-8eba76cc-b99c-44b9-9367-1c2cf659cd61 req-9207a6cf-22af-4746-8b99-ddcbfcb77ea0 service nova] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Received event network-changed-02b62aa8-27c3-43b9-a22b-eb261089a8fb {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 640.943445] env[61985]: DEBUG nova.compute.manager [req-8eba76cc-b99c-44b9-9367-1c2cf659cd61 req-9207a6cf-22af-4746-8b99-ddcbfcb77ea0 service nova] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Refreshing instance network info cache due to event network-changed-02b62aa8-27c3-43b9-a22b-eb261089a8fb. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 640.943494] env[61985]: DEBUG oslo_concurrency.lockutils [req-8eba76cc-b99c-44b9-9367-1c2cf659cd61 req-9207a6cf-22af-4746-8b99-ddcbfcb77ea0 service nova] Acquiring lock "refresh_cache-81d8927f-16fe-4823-a52f-32d213da739d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.944825] env[61985]: DEBUG oslo_concurrency.lockutils [req-8eba76cc-b99c-44b9-9367-1c2cf659cd61 req-9207a6cf-22af-4746-8b99-ddcbfcb77ea0 service nova] Acquired lock "refresh_cache-81d8927f-16fe-4823-a52f-32d213da739d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.944825] env[61985]: DEBUG nova.network.neutron [req-8eba76cc-b99c-44b9-9367-1c2cf659cd61 req-9207a6cf-22af-4746-8b99-ddcbfcb77ea0 service nova] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Refreshing network info cache for port 02b62aa8-27c3-43b9-a22b-eb261089a8fb {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 641.109045] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Releasing lock "refresh_cache-0777696f-109c-4fe1-a6d0-1498fd8e3639" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.109045] env[61985]: DEBUG nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 641.109045] env[61985]: DEBUG nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.109045] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 641.137107] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.369441] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Releasing lock "refresh_cache-8eb535d6-b993-4453-ade0-ed6c36125690" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.369994] env[61985]: DEBUG nova.compute.manager [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 641.370208] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 641.371275] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daea9631-6006-465c-9727-336907a9e9af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.381066] env[61985]: DEBUG nova.scheduler.client.report [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.384258] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 641.386257] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8924d13-9d8e-447d-a066-e3373c4d4326 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.395145] env[61985]: DEBUG oslo_vmware.api [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 641.395145] env[61985]: value = "task-935620" [ 641.395145] env[61985]: _type = "Task" [ 641.395145] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.404686] env[61985]: DEBUG oslo_vmware.api [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935620, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.468564] env[61985]: DEBUG nova.network.neutron [req-8eba76cc-b99c-44b9-9367-1c2cf659cd61 req-9207a6cf-22af-4746-8b99-ddcbfcb77ea0 service nova] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.537543] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Acquiring lock "8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.537543] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Lock "8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.616468] env[61985]: DEBUG nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 641.617050] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 641.617266] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 641.617450] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.617678] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 641.617826] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.617971] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 641.618183] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 641.618339] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 641.618521] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 641.618696] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 641.619606] env[61985]: DEBUG nova.virt.hardware [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.621120] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7682ddeb-f82f-4c46-b4be-a42b734d9f5f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.632064] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646fd8cf-00e0-4ee6-a1a3-b991a2296a52 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.647832] env[61985]: DEBUG nova.network.neutron [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.649864] env[61985]: ERROR nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6b958751-3b84-41f9-b82b-5e37e471381a, please check neutron logs for more information. [ 641.649864] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Traceback (most recent call last): [ 641.649864] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 641.649864] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] yield resources [ 641.649864] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 641.649864] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] self.driver.spawn(context, instance, image_meta, [ 641.649864] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 641.649864] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.649864] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.649864] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] vm_ref = self.build_virtual_machine(instance, [ 641.649864] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] for vif in network_info: [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] return self._sync_wrapper(fn, *args, **kwargs) [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] self.wait() [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] self[:] = self._gt.wait() [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] return self._exit_event.wait() [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 641.650269] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] current.throw(*self._exc) [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] result = function(*args, **kwargs) [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] return func(*args, **kwargs) [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] raise e [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] nwinfo = self.network_api.allocate_for_instance( [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] created_port_ids = self._update_ports_for_instance( [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] with excutils.save_and_reraise_exception(): [ 641.650745] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] self.force_reraise() [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] raise self.value [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] updated_port = self._update_port( [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] _ensure_no_port_binding_failure(port) [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] raise exception.PortBindingFailed(port_id=port['id']) [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] nova.exception.PortBindingFailed: Binding failed for port 6b958751-3b84-41f9-b82b-5e37e471381a, please check neutron logs for more information. [ 641.651176] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] [ 641.651176] env[61985]: INFO nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Terminating instance [ 641.653036] env[61985]: DEBUG nova.network.neutron [req-8eba76cc-b99c-44b9-9367-1c2cf659cd61 req-9207a6cf-22af-4746-8b99-ddcbfcb77ea0 service nova] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.654244] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Acquiring lock "refresh_cache-0947bda8-bee6-430b-bd79-eccd98d23cd8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.654403] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Acquired lock "refresh_cache-0947bda8-bee6-430b-bd79-eccd98d23cd8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.654564] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 641.886647] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.047s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.887326] env[61985]: ERROR nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 92d4dd75-b8c2-4862-9873-742bf99bc765, please check neutron logs for more information. [ 641.887326] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Traceback (most recent call last): [ 641.887326] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 641.887326] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] self.driver.spawn(context, instance, image_meta, [ 641.887326] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 641.887326] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.887326] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.887326] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] vm_ref = self.build_virtual_machine(instance, [ 641.887326] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.887326] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.887326] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] for vif in network_info: [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] return self._sync_wrapper(fn, *args, **kwargs) [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] self.wait() [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] self[:] = self._gt.wait() [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] return self._exit_event.wait() [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] result = hub.switch() [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.887862] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] return self.greenlet.switch() [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] result = function(*args, **kwargs) [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] return func(*args, **kwargs) [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] raise e [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] nwinfo = self.network_api.allocate_for_instance( [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] created_port_ids = self._update_ports_for_instance( [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] with excutils.save_and_reraise_exception(): [ 641.892224] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] self.force_reraise() [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] raise self.value [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] updated_port = self._update_port( [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] _ensure_no_port_binding_failure(port) [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] raise exception.PortBindingFailed(port_id=port['id']) [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] nova.exception.PortBindingFailed: Binding failed for port 92d4dd75-b8c2-4862-9873-742bf99bc765, please check neutron logs for more information. [ 641.892640] env[61985]: ERROR nova.compute.manager [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] [ 641.892986] env[61985]: DEBUG nova.compute.utils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Binding failed for port 92d4dd75-b8c2-4862-9873-742bf99bc765, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.892986] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.457s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.892986] env[61985]: DEBUG nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Build of instance 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba was re-scheduled: Binding failed for port 92d4dd75-b8c2-4862-9873-742bf99bc765, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 641.893346] env[61985]: DEBUG nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 641.893776] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Acquiring lock "refresh_cache-0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.893776] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Acquired lock "refresh_cache-0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.893918] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 641.905496] env[61985]: DEBUG oslo_vmware.api [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935620, 'name': PowerOffVM_Task, 'duration_secs': 0.128411} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.906322] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 641.906565] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 641.906723] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-adab890e-251c-4ff5-a7fd-704e7f58b54b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.947768] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 641.948236] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 641.949781] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Deleting the datastore file [datastore2] 8eb535d6-b993-4453-ade0-ed6c36125690 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 641.949781] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65765723-8265-447d-b0d9-666c8b215214 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.958791] env[61985]: DEBUG oslo_vmware.api [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for the task: (returnval){ [ 641.958791] env[61985]: value = "task-935622" [ 641.958791] env[61985]: _type = "Task" [ 641.958791] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.970750] env[61985]: DEBUG oslo_vmware.api [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935622, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.032274] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "2322ab72-9841-41fb-9d60-2812baabe108" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.032906] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.153517] env[61985]: INFO nova.compute.manager [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] [instance: 0777696f-109c-4fe1-a6d0-1498fd8e3639] Took 1.05 seconds to deallocate network for instance. [ 642.157058] env[61985]: DEBUG oslo_concurrency.lockutils [req-8eba76cc-b99c-44b9-9367-1c2cf659cd61 req-9207a6cf-22af-4746-8b99-ddcbfcb77ea0 service nova] Releasing lock "refresh_cache-81d8927f-16fe-4823-a52f-32d213da739d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.157058] env[61985]: DEBUG nova.compute.manager [req-8eba76cc-b99c-44b9-9367-1c2cf659cd61 req-9207a6cf-22af-4746-8b99-ddcbfcb77ea0 service nova] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Received event network-vif-deleted-02b62aa8-27c3-43b9-a22b-eb261089a8fb {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 642.172038] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.271953] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.423955] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.473688] env[61985]: DEBUG oslo_vmware.api [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Task: {'id': task-935622, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101319} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.474520] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 642.474520] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 642.474520] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 642.475160] env[61985]: INFO nova.compute.manager [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Took 1.10 seconds to destroy the instance on the hypervisor. [ 642.475254] env[61985]: DEBUG oslo.service.loopingcall [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.475419] env[61985]: DEBUG nova.compute.manager [-] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.475519] env[61985]: DEBUG nova.network.neutron [-] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 642.497258] env[61985]: DEBUG nova.network.neutron [-] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.599732] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.620251] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "4add2608-8187-42a2-9897-3e5a9049757e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.620333] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "4add2608-8187-42a2-9897-3e5a9049757e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.775124] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Releasing lock "refresh_cache-0947bda8-bee6-430b-bd79-eccd98d23cd8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.775744] env[61985]: DEBUG nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 642.776106] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-460acea3-484b-4316-abb1-286cd3f7af5f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.789264] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57196789-19e1-4a77-8162-971b93bd843d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.820650] env[61985]: WARNING nova.virt.vmwareapi.driver [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 0947bda8-bee6-430b-bd79-eccd98d23cd8 could not be found. [ 642.820770] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 642.821706] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-57e47d41-2a54-4ae6-af92-c988a22c5114 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.830632] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589b59fb-7208-470f-8ada-a428bba5b4ef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.862224] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0947bda8-bee6-430b-bd79-eccd98d23cd8 could not be found. [ 642.862224] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 642.862224] env[61985]: INFO nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Took 0.08 seconds to destroy the instance on the hypervisor. [ 642.866019] env[61985]: DEBUG oslo.service.loopingcall [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.866693] env[61985]: DEBUG nova.compute.manager [-] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.866987] env[61985]: DEBUG nova.network.neutron [-] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 642.888242] env[61985]: DEBUG nova.network.neutron [-] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.950416] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed5e5f7-22db-4708-8d95-4bc90c7d53dc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.964857] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7ba050-5eae-4677-893f-7ee2f4d85720 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.995838] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebd444b-ed2d-40d4-b714-16e0ed120780 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.001195] env[61985]: DEBUG nova.network.neutron [-] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.010106] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558c9ff5-0611-4792-b2a5-6eb19ff936f6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.025825] env[61985]: DEBUG nova.compute.provider_tree [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.085206] env[61985]: DEBUG nova.compute.manager [req-e24b4eaf-94b7-4a14-a9cb-3dab62b5f506 req-14c2ef51-2280-4c92-adc5-e503809facf2 service nova] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Received event network-changed-6b958751-3b84-41f9-b82b-5e37e471381a {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 643.085409] env[61985]: DEBUG nova.compute.manager [req-e24b4eaf-94b7-4a14-a9cb-3dab62b5f506 req-14c2ef51-2280-4c92-adc5-e503809facf2 service nova] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Refreshing instance network info cache due to event network-changed-6b958751-3b84-41f9-b82b-5e37e471381a. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 643.085634] env[61985]: DEBUG oslo_concurrency.lockutils [req-e24b4eaf-94b7-4a14-a9cb-3dab62b5f506 req-14c2ef51-2280-4c92-adc5-e503809facf2 service nova] Acquiring lock "refresh_cache-0947bda8-bee6-430b-bd79-eccd98d23cd8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.085823] env[61985]: DEBUG oslo_concurrency.lockutils [req-e24b4eaf-94b7-4a14-a9cb-3dab62b5f506 req-14c2ef51-2280-4c92-adc5-e503809facf2 service nova] Acquired lock "refresh_cache-0947bda8-bee6-430b-bd79-eccd98d23cd8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.085913] env[61985]: DEBUG nova.network.neutron [req-e24b4eaf-94b7-4a14-a9cb-3dab62b5f506 req-14c2ef51-2280-4c92-adc5-e503809facf2 service nova] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Refreshing network info cache for port 6b958751-3b84-41f9-b82b-5e37e471381a {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 643.102951] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Releasing lock "refresh_cache-0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.103254] env[61985]: DEBUG nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 643.103497] env[61985]: DEBUG nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.103577] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 643.133280] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.197842] env[61985]: INFO nova.scheduler.client.report [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Deleted allocations for instance 0777696f-109c-4fe1-a6d0-1498fd8e3639 [ 643.392999] env[61985]: DEBUG nova.network.neutron [-] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.509596] env[61985]: INFO nova.compute.manager [-] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Took 1.03 seconds to deallocate network for instance. [ 643.528973] env[61985]: DEBUG nova.scheduler.client.report [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.637158] env[61985]: DEBUG nova.network.neutron [req-e24b4eaf-94b7-4a14-a9cb-3dab62b5f506 req-14c2ef51-2280-4c92-adc5-e503809facf2 service nova] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.640370] env[61985]: DEBUG nova.network.neutron [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.710916] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dac92edd-1492-40a3-a783-331a710ef858 tempest-ServersAdminTestJSON-1646688955 tempest-ServersAdminTestJSON-1646688955-project-member] Lock "0777696f-109c-4fe1-a6d0-1498fd8e3639" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.492s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.800587] env[61985]: DEBUG nova.network.neutron [req-e24b4eaf-94b7-4a14-a9cb-3dab62b5f506 req-14c2ef51-2280-4c92-adc5-e503809facf2 service nova] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.896370] env[61985]: INFO nova.compute.manager [-] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Took 1.03 seconds to deallocate network for instance. [ 644.021793] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.037918] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.148s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.038805] env[61985]: ERROR nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4a291c97-27f0-4634-a2b4-a6e9dfb0d905, please check neutron logs for more information. [ 644.038805] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Traceback (most recent call last): [ 644.038805] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 644.038805] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] self.driver.spawn(context, instance, image_meta, [ 644.038805] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 644.038805] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.038805] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.038805] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] vm_ref = self.build_virtual_machine(instance, [ 644.038805] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.038805] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.038805] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] for vif in network_info: [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] return self._sync_wrapper(fn, *args, **kwargs) [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] self.wait() [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] self[:] = self._gt.wait() [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] return self._exit_event.wait() [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] result = hub.switch() [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 644.039483] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] return self.greenlet.switch() [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] result = function(*args, **kwargs) [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] return func(*args, **kwargs) [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] raise e [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] nwinfo = self.network_api.allocate_for_instance( [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] created_port_ids = self._update_ports_for_instance( [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] with excutils.save_and_reraise_exception(): [ 644.042053] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] self.force_reraise() [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] raise self.value [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] updated_port = self._update_port( [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] _ensure_no_port_binding_failure(port) [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] raise exception.PortBindingFailed(port_id=port['id']) [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] nova.exception.PortBindingFailed: Binding failed for port 4a291c97-27f0-4634-a2b4-a6e9dfb0d905, please check neutron logs for more information. [ 644.042840] env[61985]: ERROR nova.compute.manager [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] [ 644.043315] env[61985]: DEBUG nova.compute.utils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Binding failed for port 4a291c97-27f0-4634-a2b4-a6e9dfb0d905, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 644.043315] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.600s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.046759] env[61985]: DEBUG nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Build of instance a6dfe3a9-9745-4101-b55b-fc45c867ab27 was re-scheduled: Binding failed for port 4a291c97-27f0-4634-a2b4-a6e9dfb0d905, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 644.047594] env[61985]: DEBUG nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 644.047594] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "refresh_cache-a6dfe3a9-9745-4101-b55b-fc45c867ab27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.047721] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquired lock "refresh_cache-a6dfe3a9-9745-4101-b55b-fc45c867ab27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.047760] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 644.145068] env[61985]: INFO nova.compute.manager [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] [instance: 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba] Took 1.04 seconds to deallocate network for instance. [ 644.217106] env[61985]: DEBUG nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 644.306269] env[61985]: DEBUG oslo_concurrency.lockutils [req-e24b4eaf-94b7-4a14-a9cb-3dab62b5f506 req-14c2ef51-2280-4c92-adc5-e503809facf2 service nova] Releasing lock "refresh_cache-0947bda8-bee6-430b-bd79-eccd98d23cd8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.306627] env[61985]: DEBUG nova.compute.manager [req-e24b4eaf-94b7-4a14-a9cb-3dab62b5f506 req-14c2ef51-2280-4c92-adc5-e503809facf2 service nova] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Received event network-vif-deleted-6b958751-3b84-41f9-b82b-5e37e471381a {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 644.471526] env[61985]: INFO nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Took 0.57 seconds to detach 1 volumes for instance. [ 644.473922] env[61985]: DEBUG nova.compute.claims [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.474161] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.605763] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.747707] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.826673] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.081924] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc549c8-c248-43c2-ab19-41c7b1dee4fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.093617] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0451e35c-0312-4c7d-b02a-96cf3d05080c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.128915] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f788fd98-93e5-43d5-b578-e127ffcb7dfc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.139020] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215f7617-bef3-4796-8282-4740a9aeb5b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.152955] env[61985]: DEBUG nova.compute.provider_tree [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.182760] env[61985]: INFO nova.scheduler.client.report [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Deleted allocations for instance 0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba [ 645.330646] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Releasing lock "refresh_cache-a6dfe3a9-9745-4101-b55b-fc45c867ab27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.330646] env[61985]: DEBUG nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 645.330646] env[61985]: DEBUG nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.330646] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 645.355399] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.659164] env[61985]: DEBUG nova.scheduler.client.report [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.695293] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d74c294c-21a5-4d0b-8360-a42c857a5426 tempest-VolumesAssistedSnapshotsTest-1824634970 tempest-VolumesAssistedSnapshotsTest-1824634970-project-member] Lock "0ea7a1fd-d1bd-4de2-9e49-a2fc26f6b6ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.181s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.858531] env[61985]: DEBUG nova.network.neutron [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.163917] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.121s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.165151] env[61985]: ERROR nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b, please check neutron logs for more information. [ 646.165151] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Traceback (most recent call last): [ 646.165151] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 646.165151] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] self.driver.spawn(context, instance, image_meta, [ 646.165151] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 646.165151] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.165151] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.165151] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] vm_ref = self.build_virtual_machine(instance, [ 646.165151] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.165151] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.165151] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] for vif in network_info: [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] return self._sync_wrapper(fn, *args, **kwargs) [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] self.wait() [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] self[:] = self._gt.wait() [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] return self._exit_event.wait() [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] result = hub.switch() [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.165546] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] return self.greenlet.switch() [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] result = function(*args, **kwargs) [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] return func(*args, **kwargs) [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] raise e [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] nwinfo = self.network_api.allocate_for_instance( [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] created_port_ids = self._update_ports_for_instance( [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] with excutils.save_and_reraise_exception(): [ 646.165954] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] self.force_reraise() [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] raise self.value [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] updated_port = self._update_port( [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] _ensure_no_port_binding_failure(port) [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] raise exception.PortBindingFailed(port_id=port['id']) [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] nova.exception.PortBindingFailed: Binding failed for port 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b, please check neutron logs for more information. [ 646.166455] env[61985]: ERROR nova.compute.manager [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] [ 646.166779] env[61985]: DEBUG nova.compute.utils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Binding failed for port 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 646.168741] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.170s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.172174] env[61985]: DEBUG nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Build of instance e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c was re-scheduled: Binding failed for port 567294bd-ed64-4faa-9dfd-b6c7afd9ac4b, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 646.172655] env[61985]: DEBUG nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 646.173014] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Acquiring lock "refresh_cache-e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.173162] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Acquired lock "refresh_cache-e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.173307] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 646.198357] env[61985]: DEBUG nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.361920] env[61985]: INFO nova.compute.manager [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: a6dfe3a9-9745-4101-b55b-fc45c867ab27] Took 1.03 seconds to deallocate network for instance. [ 646.491199] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "fea0cf39-e851-409f-86f5-31cc128a44dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.491678] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.717407] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.735423] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.809246] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "794cff68-6e26-4607-96f7-eaeb41182551" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.809246] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "794cff68-6e26-4607-96f7-eaeb41182551" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.944967] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.180888] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017d3d3a-7aad-40b6-b8e5-0ee87318bf79 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.190617] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7fbefe-e76b-4ad2-ac7e-b560aa3b2d32 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.227414] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8599f1-d0fe-477d-a040-1230301f676d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.235461] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e10f07-de59-4d11-b78c-d68eb3986a62 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.249433] env[61985]: DEBUG nova.compute.provider_tree [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.411037] env[61985]: INFO nova.scheduler.client.report [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Deleted allocations for instance a6dfe3a9-9745-4101-b55b-fc45c867ab27 [ 647.447908] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Releasing lock "refresh_cache-e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.448174] env[61985]: DEBUG nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 647.448360] env[61985]: DEBUG nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.448533] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 647.502631] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.755271] env[61985]: DEBUG nova.scheduler.client.report [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.922848] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0ad818e5-5fd0-4cfe-8a47-d40d8b19a11f tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "a6dfe3a9-9745-4101-b55b-fc45c867ab27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.400s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.006966] env[61985]: DEBUG nova.network.neutron [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.262040] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.093s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.262488] env[61985]: ERROR nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a40d81ad-693f-47da-9348-398e8d746538, please check neutron logs for more information. [ 648.262488] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Traceback (most recent call last): [ 648.262488] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 648.262488] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] self.driver.spawn(context, instance, image_meta, [ 648.262488] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 648.262488] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.262488] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.262488] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] vm_ref = self.build_virtual_machine(instance, [ 648.262488] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.262488] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.262488] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] for vif in network_info: [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] return self._sync_wrapper(fn, *args, **kwargs) [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] self.wait() [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] self[:] = self._gt.wait() [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] return self._exit_event.wait() [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] result = hub.switch() [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.264278] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] return self.greenlet.switch() [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] result = function(*args, **kwargs) [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] return func(*args, **kwargs) [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] raise e [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] nwinfo = self.network_api.allocate_for_instance( [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] created_port_ids = self._update_ports_for_instance( [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] with excutils.save_and_reraise_exception(): [ 648.264847] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] self.force_reraise() [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] raise self.value [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] updated_port = self._update_port( [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] _ensure_no_port_binding_failure(port) [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] raise exception.PortBindingFailed(port_id=port['id']) [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] nova.exception.PortBindingFailed: Binding failed for port a40d81ad-693f-47da-9348-398e8d746538, please check neutron logs for more information. [ 648.265229] env[61985]: ERROR nova.compute.manager [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] [ 648.265589] env[61985]: DEBUG nova.compute.utils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Binding failed for port a40d81ad-693f-47da-9348-398e8d746538, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 648.265589] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.449s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.272597] env[61985]: DEBUG nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Build of instance 131c313b-1be0-4cac-8720-bd6ab3a0688a was re-scheduled: Binding failed for port a40d81ad-693f-47da-9348-398e8d746538, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 648.273024] env[61985]: DEBUG nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 648.273837] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquiring lock "refresh_cache-131c313b-1be0-4cac-8720-bd6ab3a0688a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.274066] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquired lock "refresh_cache-131c313b-1be0-4cac-8720-bd6ab3a0688a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.274245] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 648.426716] env[61985]: DEBUG nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.510493] env[61985]: INFO nova.compute.manager [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] [instance: e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c] Took 1.06 seconds to deallocate network for instance. [ 648.825245] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.954636] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.987394] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.289756] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7264f15-704c-4904-818a-8959b9510a35 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.301228] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d06481e-79a9-4d3f-ab02-0bd2ed873bcb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.340649] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbfd0c3-4fa6-44a2-abb3-8069fe44bc32 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.349725] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2060944b-9dab-4e7f-a24c-7ec009fc5008 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.364381] env[61985]: DEBUG nova.compute.provider_tree [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.494958] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Releasing lock "refresh_cache-131c313b-1be0-4cac-8720-bd6ab3a0688a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.495360] env[61985]: DEBUG nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 649.495441] env[61985]: DEBUG nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.495648] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 649.528808] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.567666] env[61985]: INFO nova.scheduler.client.report [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Deleted allocations for instance e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c [ 649.869174] env[61985]: DEBUG nova.scheduler.client.report [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.034332] env[61985]: DEBUG nova.network.neutron [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.080434] env[61985]: DEBUG oslo_concurrency.lockutils [None req-def49532-e4d8-49cc-b540-4c4aa4cd18a3 tempest-ServersV294TestFqdnHostnames-324054321 tempest-ServersV294TestFqdnHostnames-324054321-project-member] Lock "e7fc2f5d-7c70-4de9-bf9c-1c8f5967a86c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.341s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.378409] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.110s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.378409] env[61985]: ERROR nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ce0b4a7e-c317-4744-8740-3d762366dc34, please check neutron logs for more information. [ 650.378409] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Traceback (most recent call last): [ 650.378409] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 650.378409] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] self.driver.spawn(context, instance, image_meta, [ 650.378409] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 650.378409] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.378409] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.378409] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] vm_ref = self.build_virtual_machine(instance, [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] for vif in network_info: [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] return self._sync_wrapper(fn, *args, **kwargs) [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] self.wait() [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] self[:] = self._gt.wait() [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] return self._exit_event.wait() [ 650.378698] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] result = hub.switch() [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] return self.greenlet.switch() [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] result = function(*args, **kwargs) [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] return func(*args, **kwargs) [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] raise e [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] nwinfo = self.network_api.allocate_for_instance( [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.379072] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] created_port_ids = self._update_ports_for_instance( [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] with excutils.save_and_reraise_exception(): [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] self.force_reraise() [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] raise self.value [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] updated_port = self._update_port( [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] _ensure_no_port_binding_failure(port) [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.379452] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] raise exception.PortBindingFailed(port_id=port['id']) [ 650.379816] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] nova.exception.PortBindingFailed: Binding failed for port ce0b4a7e-c317-4744-8740-3d762366dc34, please check neutron logs for more information. [ 650.379816] env[61985]: ERROR nova.compute.manager [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] [ 650.379816] env[61985]: DEBUG nova.compute.utils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Binding failed for port ce0b4a7e-c317-4744-8740-3d762366dc34, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 650.379816] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.635s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.381150] env[61985]: INFO nova.compute.claims [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.386093] env[61985]: DEBUG nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Build of instance 5475fd04-5911-457d-85c6-30a41a8984cc was re-scheduled: Binding failed for port ce0b4a7e-c317-4744-8740-3d762366dc34, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 650.386620] env[61985]: DEBUG nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 650.387316] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquiring lock "refresh_cache-5475fd04-5911-457d-85c6-30a41a8984cc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.387316] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Acquired lock "refresh_cache-5475fd04-5911-457d-85c6-30a41a8984cc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.387316] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 650.540809] env[61985]: INFO nova.compute.manager [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 131c313b-1be0-4cac-8720-bd6ab3a0688a] Took 1.04 seconds to deallocate network for instance. [ 650.583853] env[61985]: DEBUG nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 650.915587] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.021945] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.110716] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.526762] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Releasing lock "refresh_cache-5475fd04-5911-457d-85c6-30a41a8984cc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.526762] env[61985]: DEBUG nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 651.526762] env[61985]: DEBUG nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.526762] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 651.576833] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.578345] env[61985]: INFO nova.scheduler.client.report [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Deleted allocations for instance 131c313b-1be0-4cac-8720-bd6ab3a0688a [ 651.811602] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquiring lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.812236] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.898021] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0dcec12-01ba-43d9-b399-f5be4939d3af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.907019] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a40e1c-6bd4-4b8f-9bc1-0dd1caa9f161 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.937886] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ab39f2-6ff7-4e03-875f-a22907dc71ac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.947735] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59fdc57-2f93-4777-9ead-5ade09d0ce3a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.959854] env[61985]: DEBUG nova.compute.provider_tree [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.085192] env[61985]: DEBUG nova.network.neutron [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.090949] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd3b465b-28ab-49ec-9a91-65c0e83c0441 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "131c313b-1be0-4cac-8720-bd6ab3a0688a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.916s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.462824] env[61985]: DEBUG nova.scheduler.client.report [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.589739] env[61985]: INFO nova.compute.manager [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] [instance: 5475fd04-5911-457d-85c6-30a41a8984cc] Took 1.06 seconds to deallocate network for instance. [ 652.595475] env[61985]: DEBUG nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 652.773991] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "3694e20c-ce37-4097-9991-8a06f38b2734" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.774892] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "3694e20c-ce37-4097-9991-8a06f38b2734" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.972734] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.592s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.972734] env[61985]: DEBUG nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 652.979643] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.008s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.981294] env[61985]: INFO nova.compute.claims [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.130946] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.490905] env[61985]: DEBUG nova.compute.utils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 653.493307] env[61985]: DEBUG nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 653.493494] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 653.627919] env[61985]: INFO nova.scheduler.client.report [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Deleted allocations for instance 5475fd04-5911-457d-85c6-30a41a8984cc [ 653.765580] env[61985]: DEBUG nova.policy [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c72a6ea287f241f39f6ad39d2ce3fbd5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97b7c7dc8f3b44c88ac7c552fd6cf6d6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 654.000544] env[61985]: DEBUG nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 654.140000] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cdb08110-0daa-421e-abc1-3357a8058e22 tempest-ListImageFiltersTestJSON-392564899 tempest-ListImageFiltersTestJSON-392564899-project-member] Lock "5475fd04-5911-457d-85c6-30a41a8984cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.310s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.483180] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068c5f32-c24f-4f5a-ad96-723f2b2f6c14 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.491262] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248b0247-6542-4f2d-93ab-609dad77e2b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.530368] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a35cf0-2f88-46e0-9622-3e8081d312c9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.542587] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68aa8fef-6c5c-47f1-90d8-a4411ec7222f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.556664] env[61985]: DEBUG nova.compute.provider_tree [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.638904] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Successfully created port: 36457600-3fc2-43e3-a354-7f28da7c3e44 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.646535] env[61985]: DEBUG nova.compute.manager [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 654.738488] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "f8bc8b71-0317-479d-b2f9-9471757f0774" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.738848] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "f8bc8b71-0317-479d-b2f9-9471757f0774" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.038988] env[61985]: DEBUG nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 655.063798] env[61985]: DEBUG nova.scheduler.client.report [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.080697] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:24:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='116299873',id=20,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-2054099518',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 655.080957] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 655.081129] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 655.081308] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 655.081445] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 655.081584] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 655.081781] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 655.081933] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 655.082104] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 655.084290] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 655.084290] env[61985]: DEBUG nova.virt.hardware [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.084290] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cffa17-7b21-4ce1-bb0a-3311f6185118 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.092497] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e33c4a4-76a1-4fea-95d1-cae905062870 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.171808] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.566476] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.566660] env[61985]: DEBUG nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 655.574579] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.298s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.075536] env[61985]: DEBUG nova.compute.utils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 656.077877] env[61985]: DEBUG nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 656.077877] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 656.233047] env[61985]: DEBUG nova.policy [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16c90171651d4790a1f6f706b29f4b90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97a294a9bd8444c480ac4762c1cc5312', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 656.531918] env[61985]: DEBUG nova.compute.manager [req-d8ce5de8-f8c4-4bd5-8519-2c09a6620150 req-1ddcac54-eb39-4489-b241-331b7b94b6d9 service nova] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Received event network-changed-36457600-3fc2-43e3-a354-7f28da7c3e44 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 656.532139] env[61985]: DEBUG nova.compute.manager [req-d8ce5de8-f8c4-4bd5-8519-2c09a6620150 req-1ddcac54-eb39-4489-b241-331b7b94b6d9 service nova] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Refreshing instance network info cache due to event network-changed-36457600-3fc2-43e3-a354-7f28da7c3e44. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 656.532354] env[61985]: DEBUG oslo_concurrency.lockutils [req-d8ce5de8-f8c4-4bd5-8519-2c09a6620150 req-1ddcac54-eb39-4489-b241-331b7b94b6d9 service nova] Acquiring lock "refresh_cache-9ac9f4b8-fcef-458a-913c-72cfc3322b6e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.532494] env[61985]: DEBUG oslo_concurrency.lockutils [req-d8ce5de8-f8c4-4bd5-8519-2c09a6620150 req-1ddcac54-eb39-4489-b241-331b7b94b6d9 service nova] Acquired lock "refresh_cache-9ac9f4b8-fcef-458a-913c-72cfc3322b6e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.532650] env[61985]: DEBUG nova.network.neutron [req-d8ce5de8-f8c4-4bd5-8519-2c09a6620150 req-1ddcac54-eb39-4489-b241-331b7b94b6d9 service nova] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Refreshing network info cache for port 36457600-3fc2-43e3-a354-7f28da7c3e44 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 656.583145] env[61985]: DEBUG nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 656.595993] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0f1132-d7a4-40ed-970e-df0d9f0c2921 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.604263] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0eb412f-5339-47a0-8cb8-71314da8be86 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.645248] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb83202-f7e7-476a-a955-eb7ccdaa0459 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.657062] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11118c34-ee50-4e14-9166-19b72db64f98 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.681229] env[61985]: DEBUG nova.compute.provider_tree [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.991509] env[61985]: ERROR nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 36457600-3fc2-43e3-a354-7f28da7c3e44, please check neutron logs for more information. [ 656.991509] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.991509] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.991509] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.991509] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.991509] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.991509] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.991509] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.991509] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.991509] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 656.991509] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.991509] env[61985]: ERROR nova.compute.manager raise self.value [ 656.991509] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.991509] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.991509] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.991509] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.992438] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.992438] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.992438] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 36457600-3fc2-43e3-a354-7f28da7c3e44, please check neutron logs for more information. [ 656.992438] env[61985]: ERROR nova.compute.manager [ 656.992438] env[61985]: Traceback (most recent call last): [ 656.992438] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.992438] env[61985]: listener.cb(fileno) [ 656.992438] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.992438] env[61985]: result = function(*args, **kwargs) [ 656.992438] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.992438] env[61985]: return func(*args, **kwargs) [ 656.992438] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.992438] env[61985]: raise e [ 656.992438] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.992438] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 656.992438] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.992438] env[61985]: created_port_ids = self._update_ports_for_instance( [ 656.992438] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.992438] env[61985]: with excutils.save_and_reraise_exception(): [ 656.992438] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.992438] env[61985]: self.force_reraise() [ 656.992438] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.992438] env[61985]: raise self.value [ 656.992438] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.992438] env[61985]: updated_port = self._update_port( [ 656.992438] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.992438] env[61985]: _ensure_no_port_binding_failure(port) [ 656.992438] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.992438] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.994041] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 36457600-3fc2-43e3-a354-7f28da7c3e44, please check neutron logs for more information. [ 656.994041] env[61985]: Removing descriptor: 18 [ 656.994041] env[61985]: ERROR nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 36457600-3fc2-43e3-a354-7f28da7c3e44, please check neutron logs for more information. [ 656.994041] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Traceback (most recent call last): [ 656.994041] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 656.994041] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] yield resources [ 656.994041] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 656.994041] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] self.driver.spawn(context, instance, image_meta, [ 656.994041] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 656.994041] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.994041] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.994041] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] vm_ref = self.build_virtual_machine(instance, [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] for vif in network_info: [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] return self._sync_wrapper(fn, *args, **kwargs) [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] self.wait() [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] self[:] = self._gt.wait() [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] return self._exit_event.wait() [ 656.994669] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] result = hub.switch() [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] return self.greenlet.switch() [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] result = function(*args, **kwargs) [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] return func(*args, **kwargs) [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] raise e [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] nwinfo = self.network_api.allocate_for_instance( [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.002434] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] created_port_ids = self._update_ports_for_instance( [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] with excutils.save_and_reraise_exception(): [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] self.force_reraise() [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] raise self.value [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] updated_port = self._update_port( [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] _ensure_no_port_binding_failure(port) [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.003242] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] raise exception.PortBindingFailed(port_id=port['id']) [ 657.003750] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] nova.exception.PortBindingFailed: Binding failed for port 36457600-3fc2-43e3-a354-7f28da7c3e44, please check neutron logs for more information. [ 657.003750] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] [ 657.003750] env[61985]: INFO nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Terminating instance [ 657.003750] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Acquiring lock "refresh_cache-9ac9f4b8-fcef-458a-913c-72cfc3322b6e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.085937] env[61985]: DEBUG nova.network.neutron [req-d8ce5de8-f8c4-4bd5-8519-2c09a6620150 req-1ddcac54-eb39-4489-b241-331b7b94b6d9 service nova] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.185484] env[61985]: DEBUG nova.scheduler.client.report [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.485747] env[61985]: DEBUG nova.network.neutron [req-d8ce5de8-f8c4-4bd5-8519-2c09a6620150 req-1ddcac54-eb39-4489-b241-331b7b94b6d9 service nova] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.490677] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Successfully created port: 8d895f96-2327-47ba-a747-e6b97832e9fc {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 657.595582] env[61985]: DEBUG nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 657.627868] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.628167] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.628378] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.628601] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.628766] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.628917] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.629412] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.629629] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.629859] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.630046] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.630419] env[61985]: DEBUG nova.virt.hardware [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.632168] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f207c52c-0747-4f2b-972e-90bd2f9308c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.642275] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735653b8-aab7-4c61-b119-201604b3a32e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.691246] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.118s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.691865] env[61985]: ERROR nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 02b62aa8-27c3-43b9-a22b-eb261089a8fb, please check neutron logs for more information. [ 657.691865] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Traceback (most recent call last): [ 657.691865] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 657.691865] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] self.driver.spawn(context, instance, image_meta, [ 657.691865] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 657.691865] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.691865] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.691865] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] vm_ref = self.build_virtual_machine(instance, [ 657.691865] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.691865] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.691865] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] for vif in network_info: [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] return self._sync_wrapper(fn, *args, **kwargs) [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] self.wait() [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] self[:] = self._gt.wait() [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] return self._exit_event.wait() [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] result = hub.switch() [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.692218] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] return self.greenlet.switch() [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] result = function(*args, **kwargs) [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] return func(*args, **kwargs) [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] raise e [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] nwinfo = self.network_api.allocate_for_instance( [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] created_port_ids = self._update_ports_for_instance( [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] with excutils.save_and_reraise_exception(): [ 657.692582] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] self.force_reraise() [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] raise self.value [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] updated_port = self._update_port( [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] _ensure_no_port_binding_failure(port) [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] raise exception.PortBindingFailed(port_id=port['id']) [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] nova.exception.PortBindingFailed: Binding failed for port 02b62aa8-27c3-43b9-a22b-eb261089a8fb, please check neutron logs for more information. [ 657.692935] env[61985]: ERROR nova.compute.manager [instance: 81d8927f-16fe-4823-a52f-32d213da739d] [ 657.693378] env[61985]: DEBUG nova.compute.utils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Binding failed for port 02b62aa8-27c3-43b9-a22b-eb261089a8fb, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 657.693870] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.672s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.694214] env[61985]: DEBUG nova.objects.instance [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lazy-loading 'resources' on Instance uuid 8eb535d6-b993-4453-ade0-ed6c36125690 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 657.695501] env[61985]: DEBUG nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Build of instance 81d8927f-16fe-4823-a52f-32d213da739d was re-scheduled: Binding failed for port 02b62aa8-27c3-43b9-a22b-eb261089a8fb, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 657.696012] env[61985]: DEBUG nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 657.696250] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "refresh_cache-81d8927f-16fe-4823-a52f-32d213da739d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.696399] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquired lock "refresh_cache-81d8927f-16fe-4823-a52f-32d213da739d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.696559] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.989150] env[61985]: DEBUG oslo_concurrency.lockutils [req-d8ce5de8-f8c4-4bd5-8519-2c09a6620150 req-1ddcac54-eb39-4489-b241-331b7b94b6d9 service nova] Releasing lock "refresh_cache-9ac9f4b8-fcef-458a-913c-72cfc3322b6e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.989554] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Acquired lock "refresh_cache-9ac9f4b8-fcef-458a-913c-72cfc3322b6e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.989745] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 658.249538] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.474737] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.537347] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.662351] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f828cba8-9cff-4305-b039-4e97f7e8ff8a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.673936] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e615e284-6583-4fdf-a222-fe8e0f251e1e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.704880] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ce1435-4f44-4ca0-a415-68231f818fbb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.712929] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111aa62d-8939-437b-93bf-c798754a3520 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.726909] env[61985]: DEBUG nova.compute.provider_tree [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.801756] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.966773] env[61985]: DEBUG nova.compute.manager [req-0d2acdf4-9ccf-41e3-9574-aa051e82d869 req-a883bcf3-e4ea-4563-a0dd-0a2ecc0275ce service nova] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Received event network-vif-deleted-36457600-3fc2-43e3-a354-7f28da7c3e44 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 658.982615] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Releasing lock "refresh_cache-81d8927f-16fe-4823-a52f-32d213da739d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.982854] env[61985]: DEBUG nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 658.983060] env[61985]: DEBUG nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.983251] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 659.020712] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.230383] env[61985]: DEBUG nova.scheduler.client.report [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.305099] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Releasing lock "refresh_cache-9ac9f4b8-fcef-458a-913c-72cfc3322b6e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.305617] env[61985]: DEBUG nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 659.305823] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 659.306130] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-273d4b9f-b3d8-42a2-a247-b523411c30e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.317165] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2aa49b-528c-4877-931c-5dafc2d526f3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.340532] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9ac9f4b8-fcef-458a-913c-72cfc3322b6e could not be found. [ 659.343080] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 659.343080] env[61985]: INFO nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 659.343080] env[61985]: DEBUG oslo.service.loopingcall [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.343080] env[61985]: DEBUG nova.compute.manager [-] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.343080] env[61985]: DEBUG nova.network.neutron [-] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 659.377811] env[61985]: DEBUG nova.network.neutron [-] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.522268] env[61985]: DEBUG nova.network.neutron [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.736377] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.042s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.739645] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.265s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.766347] env[61985]: INFO nova.scheduler.client.report [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Deleted allocations for instance 8eb535d6-b993-4453-ade0-ed6c36125690 [ 659.877759] env[61985]: DEBUG nova.network.neutron [-] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.025800] env[61985]: INFO nova.compute.manager [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 81d8927f-16fe-4823-a52f-32d213da739d] Took 1.04 seconds to deallocate network for instance. [ 660.078077] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 660.078077] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 660.131568] env[61985]: ERROR nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8d895f96-2327-47ba-a747-e6b97832e9fc, please check neutron logs for more information. [ 660.131568] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.131568] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.131568] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.131568] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.131568] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.131568] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.131568] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.131568] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.131568] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 660.131568] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.131568] env[61985]: ERROR nova.compute.manager raise self.value [ 660.131568] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.131568] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.131568] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.131568] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.132081] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.132081] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.132081] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8d895f96-2327-47ba-a747-e6b97832e9fc, please check neutron logs for more information. [ 660.132081] env[61985]: ERROR nova.compute.manager [ 660.132081] env[61985]: Traceback (most recent call last): [ 660.132081] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.132081] env[61985]: listener.cb(fileno) [ 660.132081] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.132081] env[61985]: result = function(*args, **kwargs) [ 660.132081] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.132081] env[61985]: return func(*args, **kwargs) [ 660.132081] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.132081] env[61985]: raise e [ 660.132081] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.132081] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 660.132081] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.132081] env[61985]: created_port_ids = self._update_ports_for_instance( [ 660.132081] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.132081] env[61985]: with excutils.save_and_reraise_exception(): [ 660.132081] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.132081] env[61985]: self.force_reraise() [ 660.132081] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.132081] env[61985]: raise self.value [ 660.132081] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.132081] env[61985]: updated_port = self._update_port( [ 660.132081] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.132081] env[61985]: _ensure_no_port_binding_failure(port) [ 660.132081] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.132081] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.133142] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 8d895f96-2327-47ba-a747-e6b97832e9fc, please check neutron logs for more information. [ 660.133142] env[61985]: Removing descriptor: 17 [ 660.133142] env[61985]: ERROR nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8d895f96-2327-47ba-a747-e6b97832e9fc, please check neutron logs for more information. [ 660.133142] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Traceback (most recent call last): [ 660.133142] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 660.133142] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] yield resources [ 660.133142] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 660.133142] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] self.driver.spawn(context, instance, image_meta, [ 660.133142] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 660.133142] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.133142] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.133142] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] vm_ref = self.build_virtual_machine(instance, [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] for vif in network_info: [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] return self._sync_wrapper(fn, *args, **kwargs) [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] self.wait() [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] self[:] = self._gt.wait() [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] return self._exit_event.wait() [ 660.133606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] result = hub.switch() [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] return self.greenlet.switch() [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] result = function(*args, **kwargs) [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] return func(*args, **kwargs) [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] raise e [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] nwinfo = self.network_api.allocate_for_instance( [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.133970] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] created_port_ids = self._update_ports_for_instance( [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] with excutils.save_and_reraise_exception(): [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] self.force_reraise() [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] raise self.value [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] updated_port = self._update_port( [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] _ensure_no_port_binding_failure(port) [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.134368] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] raise exception.PortBindingFailed(port_id=port['id']) [ 660.134829] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] nova.exception.PortBindingFailed: Binding failed for port 8d895f96-2327-47ba-a747-e6b97832e9fc, please check neutron logs for more information. [ 660.134829] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] [ 660.134829] env[61985]: INFO nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Terminating instance [ 660.139335] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquiring lock "refresh_cache-7f32151e-7be0-40e2-b210-72765b7551bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.139507] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquired lock "refresh_cache-7f32151e-7be0-40e2-b210-72765b7551bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.139679] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 660.286307] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c3797ade-5af2-447c-a9b2-11d710dd34cc tempest-ServerDiagnosticsV248Test-552807454 tempest-ServerDiagnosticsV248Test-552807454-project-member] Lock "8eb535d6-b993-4453-ade0-ed6c36125690" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.095s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.383427] env[61985]: INFO nova.compute.manager [-] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Took 1.04 seconds to deallocate network for instance. [ 660.389749] env[61985]: DEBUG nova.compute.claims [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 660.389993] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.589310] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 660.589475] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 660.590237] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Rebuilding the list of instances to heal {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9936}} [ 660.672130] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.693652] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "b659ff28-5101-4825-84ea-111351c81145" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.693896] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "b659ff28-5101-4825-84ea-111351c81145" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.776012] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.813534] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e5bfa2-4a3d-47bc-a7a7-bb2350ffaff3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.821396] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a224576-4e71-4d2b-a160-aa8287331e98 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.854282] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9e813a-428f-4fb1-b278-7b865f16efa2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.862771] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3dbff5-df13-4d52-9c74-8bb7304f8d4a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.876544] env[61985]: DEBUG nova.compute.provider_tree [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.067255] env[61985]: INFO nova.scheduler.client.report [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Deleted allocations for instance 81d8927f-16fe-4823-a52f-32d213da739d [ 661.092640] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 661.092863] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 661.092999] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Didn't find any instances for network info cache update. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10018}} [ 661.093996] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 661.093996] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 661.093996] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 661.093996] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 661.095198] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 661.095600] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 661.096397] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 661.097425] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 661.276042] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.277670] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.278035] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Releasing lock "refresh_cache-7f32151e-7be0-40e2-b210-72765b7551bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.278909] env[61985]: DEBUG nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 661.279329] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 661.279937] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ecaba683-e775-4e7c-b0de-05d6351f3dbd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.289282] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6375fc37-b161-4c57-ac89-572d4f89a426 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.315761] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7f32151e-7be0-40e2-b210-72765b7551bc could not be found. [ 661.315879] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 661.318042] env[61985]: INFO nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 661.318042] env[61985]: DEBUG oslo.service.loopingcall [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 661.318042] env[61985]: DEBUG nova.compute.manager [-] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.318042] env[61985]: DEBUG nova.network.neutron [-] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 661.338923] env[61985]: DEBUG nova.network.neutron [-] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.382697] env[61985]: DEBUG nova.scheduler.client.report [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.403586] env[61985]: DEBUG nova.compute.manager [req-d0058133-49f5-4bb1-a2b2-7d9ccf55d0e9 req-f7d37f1f-2c08-437d-b6e9-8c63a3090aea service nova] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Received event network-changed-8d895f96-2327-47ba-a747-e6b97832e9fc {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 661.403586] env[61985]: DEBUG nova.compute.manager [req-d0058133-49f5-4bb1-a2b2-7d9ccf55d0e9 req-f7d37f1f-2c08-437d-b6e9-8c63a3090aea service nova] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Refreshing instance network info cache due to event network-changed-8d895f96-2327-47ba-a747-e6b97832e9fc. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 661.403586] env[61985]: DEBUG oslo_concurrency.lockutils [req-d0058133-49f5-4bb1-a2b2-7d9ccf55d0e9 req-f7d37f1f-2c08-437d-b6e9-8c63a3090aea service nova] Acquiring lock "refresh_cache-7f32151e-7be0-40e2-b210-72765b7551bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.403586] env[61985]: DEBUG oslo_concurrency.lockutils [req-d0058133-49f5-4bb1-a2b2-7d9ccf55d0e9 req-f7d37f1f-2c08-437d-b6e9-8c63a3090aea service nova] Acquired lock "refresh_cache-7f32151e-7be0-40e2-b210-72765b7551bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.403586] env[61985]: DEBUG nova.network.neutron [req-d0058133-49f5-4bb1-a2b2-7d9ccf55d0e9 req-f7d37f1f-2c08-437d-b6e9-8c63a3090aea service nova] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Refreshing network info cache for port 8d895f96-2327-47ba-a747-e6b97832e9fc {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 661.577113] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cd11a7b9-7c0a-419f-92e6-0a90ac6bec93 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "81d8927f-16fe-4823-a52f-32d213da739d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.171s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.601349] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.841411] env[61985]: DEBUG nova.network.neutron [-] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.890199] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.150s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.890199] env[61985]: ERROR nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6b958751-3b84-41f9-b82b-5e37e471381a, please check neutron logs for more information. [ 661.890199] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Traceback (most recent call last): [ 661.890199] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 661.890199] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] self.driver.spawn(context, instance, image_meta, [ 661.890199] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.890199] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.890199] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.890199] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] vm_ref = self.build_virtual_machine(instance, [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] for vif in network_info: [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] return self._sync_wrapper(fn, *args, **kwargs) [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] self.wait() [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] self[:] = self._gt.wait() [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] return self._exit_event.wait() [ 661.890608] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] current.throw(*self._exc) [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] result = function(*args, **kwargs) [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] return func(*args, **kwargs) [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] raise e [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] nwinfo = self.network_api.allocate_for_instance( [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] created_port_ids = self._update_ports_for_instance( [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.891040] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] with excutils.save_and_reraise_exception(): [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] self.force_reraise() [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] raise self.value [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] updated_port = self._update_port( [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] _ensure_no_port_binding_failure(port) [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] raise exception.PortBindingFailed(port_id=port['id']) [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] nova.exception.PortBindingFailed: Binding failed for port 6b958751-3b84-41f9-b82b-5e37e471381a, please check neutron logs for more information. [ 661.891440] env[61985]: ERROR nova.compute.manager [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] [ 661.891840] env[61985]: DEBUG nova.compute.utils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Binding failed for port 6b958751-3b84-41f9-b82b-5e37e471381a, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 661.891840] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.145s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.894368] env[61985]: INFO nova.compute.claims [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 661.898522] env[61985]: DEBUG nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Build of instance 0947bda8-bee6-430b-bd79-eccd98d23cd8 was re-scheduled: Binding failed for port 6b958751-3b84-41f9-b82b-5e37e471381a, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 661.902019] env[61985]: DEBUG nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 661.902019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Acquiring lock "refresh_cache-0947bda8-bee6-430b-bd79-eccd98d23cd8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.902019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Acquired lock "refresh_cache-0947bda8-bee6-430b-bd79-eccd98d23cd8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.902019] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.939451] env[61985]: DEBUG nova.network.neutron [req-d0058133-49f5-4bb1-a2b2-7d9ccf55d0e9 req-f7d37f1f-2c08-437d-b6e9-8c63a3090aea service nova] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.044596] env[61985]: DEBUG nova.network.neutron [req-d0058133-49f5-4bb1-a2b2-7d9ccf55d0e9 req-f7d37f1f-2c08-437d-b6e9-8c63a3090aea service nova] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.080396] env[61985]: DEBUG nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 662.346200] env[61985]: INFO nova.compute.manager [-] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Took 1.03 seconds to deallocate network for instance. [ 662.346790] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquiring lock "4aca0ecb-4ae6-4400-accd-d71782b2806d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.347021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lock "4aca0ecb-4ae6-4400-accd-d71782b2806d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.349021] env[61985]: DEBUG nova.compute.claims [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 662.349271] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.424042] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.552027] env[61985]: DEBUG oslo_concurrency.lockutils [req-d0058133-49f5-4bb1-a2b2-7d9ccf55d0e9 req-f7d37f1f-2c08-437d-b6e9-8c63a3090aea service nova] Releasing lock "refresh_cache-7f32151e-7be0-40e2-b210-72765b7551bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.552027] env[61985]: DEBUG nova.compute.manager [req-d0058133-49f5-4bb1-a2b2-7d9ccf55d0e9 req-f7d37f1f-2c08-437d-b6e9-8c63a3090aea service nova] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Received event network-vif-deleted-8d895f96-2327-47ba-a747-e6b97832e9fc {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 662.611565] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.648402] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.811170] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.811692] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.155856] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Releasing lock "refresh_cache-0947bda8-bee6-430b-bd79-eccd98d23cd8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.155856] env[61985]: DEBUG nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 663.155856] env[61985]: DEBUG nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.156085] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.171470] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.452362] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4676ad3d-4f61-435b-8671-71e19c147fed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.462425] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e23c8e-ea5b-48c2-98e4-e4570e280cd9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.503687] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9055038-dee6-414e-b981-0863198d131e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.512418] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048bd103-463d-4f24-9366-17d1bf36cee2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.533525] env[61985]: DEBUG nova.compute.provider_tree [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.676214] env[61985]: DEBUG nova.network.neutron [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.037672] env[61985]: DEBUG nova.scheduler.client.report [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.178416] env[61985]: INFO nova.compute.manager [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] [instance: 0947bda8-bee6-430b-bd79-eccd98d23cd8] Took 1.02 seconds to deallocate network for instance. [ 664.551017] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.657s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.551017] env[61985]: DEBUG nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 664.558863] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.823s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.561963] env[61985]: INFO nova.compute.claims [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.719067] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "8f6cd002-b3c2-4276-b195-15d09a143d31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.719450] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8f6cd002-b3c2-4276-b195-15d09a143d31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.063025] env[61985]: DEBUG nova.compute.utils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 665.063025] env[61985]: DEBUG nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 665.063025] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 665.230014] env[61985]: INFO nova.scheduler.client.report [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Deleted allocations for instance 0947bda8-bee6-430b-bd79-eccd98d23cd8 [ 665.320034] env[61985]: DEBUG nova.policy [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2dba42e753b4641862b3a03ce563892', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff085eb00ec4486b923bcdfd67406b14', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 665.567927] env[61985]: DEBUG nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 665.736243] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ca49a3b1-a328-4acd-a7a7-a2d4c38e426c tempest-ServersTestBootFromVolume-832498689 tempest-ServersTestBootFromVolume-832498689-project-member] Lock "0947bda8-bee6-430b-bd79-eccd98d23cd8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.405s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.790148] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Successfully created port: 892a5f88-22a3-44b3-8b77-8b48f536dd80 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 666.111408] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a2dd1d-2254-484c-aea4-7c51f7ebf9be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.120129] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ff58ff-5791-4545-a9ad-999e2df6e295 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.156181] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9493ec7c-6b87-41de-a599-e0bc368154b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.164492] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1307536f-94a0-4956-a901-04576b0d37e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.184081] env[61985]: DEBUG nova.compute.provider_tree [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.244441] env[61985]: DEBUG nova.compute.manager [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 666.580285] env[61985]: DEBUG nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 666.625224] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.625871] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.626105] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.626344] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.626503] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.626655] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.626874] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.627042] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.627259] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.627375] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.627623] env[61985]: DEBUG nova.virt.hardware [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.628532] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd182234-0749-4c20-bba4-5ad19df2453f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.640609] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d40153-907f-41b6-b772-8802e1d0c5c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.689019] env[61985]: DEBUG nova.scheduler.client.report [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.771226] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.949758] env[61985]: DEBUG nova.compute.manager [req-37d1822b-eba9-43cc-a20c-deef09550bbf req-36bbe876-e8e7-4805-acc7-5866743c45ff service nova] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Received event network-changed-892a5f88-22a3-44b3-8b77-8b48f536dd80 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 666.949947] env[61985]: DEBUG nova.compute.manager [req-37d1822b-eba9-43cc-a20c-deef09550bbf req-36bbe876-e8e7-4805-acc7-5866743c45ff service nova] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Refreshing instance network info cache due to event network-changed-892a5f88-22a3-44b3-8b77-8b48f536dd80. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 666.950322] env[61985]: DEBUG oslo_concurrency.lockutils [req-37d1822b-eba9-43cc-a20c-deef09550bbf req-36bbe876-e8e7-4805-acc7-5866743c45ff service nova] Acquiring lock "refresh_cache-535a5c48-c816-44f7-a58c-734878d90cc7" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.951011] env[61985]: DEBUG oslo_concurrency.lockutils [req-37d1822b-eba9-43cc-a20c-deef09550bbf req-36bbe876-e8e7-4805-acc7-5866743c45ff service nova] Acquired lock "refresh_cache-535a5c48-c816-44f7-a58c-734878d90cc7" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.951011] env[61985]: DEBUG nova.network.neutron [req-37d1822b-eba9-43cc-a20c-deef09550bbf req-36bbe876-e8e7-4805-acc7-5866743c45ff service nova] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Refreshing network info cache for port 892a5f88-22a3-44b3-8b77-8b48f536dd80 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 667.194110] env[61985]: ERROR nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 892a5f88-22a3-44b3-8b77-8b48f536dd80, please check neutron logs for more information. [ 667.194110] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 667.194110] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.194110] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 667.194110] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.194110] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 667.194110] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.194110] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 667.194110] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.194110] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 667.194110] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.194110] env[61985]: ERROR nova.compute.manager raise self.value [ 667.194110] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.194110] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 667.194110] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.194110] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 667.194526] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.194526] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 667.194526] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 892a5f88-22a3-44b3-8b77-8b48f536dd80, please check neutron logs for more information. [ 667.194526] env[61985]: ERROR nova.compute.manager [ 667.194526] env[61985]: Traceback (most recent call last): [ 667.194526] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 667.194526] env[61985]: listener.cb(fileno) [ 667.194526] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.194526] env[61985]: result = function(*args, **kwargs) [ 667.194526] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.194526] env[61985]: return func(*args, **kwargs) [ 667.194526] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.194526] env[61985]: raise e [ 667.194526] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.194526] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 667.194526] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.194526] env[61985]: created_port_ids = self._update_ports_for_instance( [ 667.194526] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.194526] env[61985]: with excutils.save_and_reraise_exception(): [ 667.194526] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.194526] env[61985]: self.force_reraise() [ 667.194526] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.194526] env[61985]: raise self.value [ 667.194526] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.194526] env[61985]: updated_port = self._update_port( [ 667.194526] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.194526] env[61985]: _ensure_no_port_binding_failure(port) [ 667.194526] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.194526] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 667.195101] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 892a5f88-22a3-44b3-8b77-8b48f536dd80, please check neutron logs for more information. [ 667.195101] env[61985]: Removing descriptor: 17 [ 667.195275] env[61985]: ERROR nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 892a5f88-22a3-44b3-8b77-8b48f536dd80, please check neutron logs for more information. [ 667.195275] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Traceback (most recent call last): [ 667.195275] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 667.195275] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] yield resources [ 667.195275] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 667.195275] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] self.driver.spawn(context, instance, image_meta, [ 667.195275] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 667.195275] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.195275] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.195275] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] vm_ref = self.build_virtual_machine(instance, [ 667.195275] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] for vif in network_info: [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] return self._sync_wrapper(fn, *args, **kwargs) [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] self.wait() [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] self[:] = self._gt.wait() [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] return self._exit_event.wait() [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.195503] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] result = hub.switch() [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] return self.greenlet.switch() [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] result = function(*args, **kwargs) [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] return func(*args, **kwargs) [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] raise e [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] nwinfo = self.network_api.allocate_for_instance( [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] created_port_ids = self._update_ports_for_instance( [ 667.195754] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] with excutils.save_and_reraise_exception(): [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] self.force_reraise() [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] raise self.value [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] updated_port = self._update_port( [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] _ensure_no_port_binding_failure(port) [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] raise exception.PortBindingFailed(port_id=port['id']) [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] nova.exception.PortBindingFailed: Binding failed for port 892a5f88-22a3-44b3-8b77-8b48f536dd80, please check neutron logs for more information. [ 667.196036] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] [ 667.196370] env[61985]: INFO nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Terminating instance [ 667.198304] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.640s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.198816] env[61985]: DEBUG nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 667.201906] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "refresh_cache-535a5c48-c816-44f7-a58c-734878d90cc7" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.202184] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.248s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.204041] env[61985]: INFO nova.compute.claims [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.483250] env[61985]: DEBUG nova.network.neutron [req-37d1822b-eba9-43cc-a20c-deef09550bbf req-36bbe876-e8e7-4805-acc7-5866743c45ff service nova] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.648270] env[61985]: DEBUG nova.network.neutron [req-37d1822b-eba9-43cc-a20c-deef09550bbf req-36bbe876-e8e7-4805-acc7-5866743c45ff service nova] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.712053] env[61985]: DEBUG nova.compute.utils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.716036] env[61985]: DEBUG nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 667.716505] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 667.784138] env[61985]: DEBUG nova.policy [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da7c7d5d10b34b4ebdede5ff3cfd8b01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e66b7a6e7354b06a77295eadf4e5e0f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 668.105156] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquiring lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.105156] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.155033] env[61985]: DEBUG oslo_concurrency.lockutils [req-37d1822b-eba9-43cc-a20c-deef09550bbf req-36bbe876-e8e7-4805-acc7-5866743c45ff service nova] Releasing lock "refresh_cache-535a5c48-c816-44f7-a58c-734878d90cc7" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.155486] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquired lock "refresh_cache-535a5c48-c816-44f7-a58c-734878d90cc7" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.156015] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 668.223017] env[61985]: DEBUG nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 668.320094] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Successfully created port: a1fe1a7b-3745-42c6-9f72-0afc45077ddf {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.692575] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.774478] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e294bcd5-cee1-4269-a0c7-7bc9c734ca64 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.790021] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc43d4f1-f1c9-46f3-91a2-14f9258af04f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.822032] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cc1217-8890-4933-94db-a48c0ce10160 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.829558] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113d4302-d54f-46e7-9def-35d018f1a087 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.843383] env[61985]: DEBUG nova.compute.provider_tree [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.848844] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.046298] env[61985]: DEBUG nova.compute.manager [req-ed76b9d8-5468-47df-8daf-c5ba93807ca5 req-8dc65713-cabe-47d0-bc43-c4df967df680 service nova] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Received event network-vif-deleted-892a5f88-22a3-44b3-8b77-8b48f536dd80 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 669.239153] env[61985]: DEBUG nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 669.269905] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.270452] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.270689] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.270918] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.271564] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.271831] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.272091] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.272823] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.273131] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.273392] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.273591] env[61985]: DEBUG nova.virt.hardware [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.274461] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22feac0e-5c4a-48e2-af65-3c700b03dbf0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.283070] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d0d37a-c458-4cac-8a7c-2a20ff5d9191 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.347393] env[61985]: DEBUG nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.352443] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Releasing lock "refresh_cache-535a5c48-c816-44f7-a58c-734878d90cc7" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.353048] env[61985]: DEBUG nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 669.353413] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 669.354471] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67f60995-fbd7-4032-9820-4572c60c3320 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.368107] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc5a9af-6482-4f63-a732-8b4e35baf6a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.401447] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 535a5c48-c816-44f7-a58c-734878d90cc7 could not be found. [ 669.401600] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 669.402197] env[61985]: INFO nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Took 0.05 seconds to destroy the instance on the hypervisor. [ 669.402197] env[61985]: DEBUG oslo.service.loopingcall [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.402316] env[61985]: DEBUG nova.compute.manager [-] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.402441] env[61985]: DEBUG nova.network.neutron [-] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 669.427657] env[61985]: DEBUG nova.network.neutron [-] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.855989] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.654s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.856566] env[61985]: DEBUG nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 669.860027] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.749s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.861505] env[61985]: INFO nova.compute.claims [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 669.928905] env[61985]: ERROR nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a1fe1a7b-3745-42c6-9f72-0afc45077ddf, please check neutron logs for more information. [ 669.928905] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.928905] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.928905] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.928905] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.928905] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.928905] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.928905] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.928905] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.928905] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 669.928905] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.928905] env[61985]: ERROR nova.compute.manager raise self.value [ 669.928905] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.928905] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.928905] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.928905] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.929454] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.929454] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.929454] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a1fe1a7b-3745-42c6-9f72-0afc45077ddf, please check neutron logs for more information. [ 669.929454] env[61985]: ERROR nova.compute.manager [ 669.929454] env[61985]: Traceback (most recent call last): [ 669.929454] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.929454] env[61985]: listener.cb(fileno) [ 669.929454] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.929454] env[61985]: result = function(*args, **kwargs) [ 669.929454] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.929454] env[61985]: return func(*args, **kwargs) [ 669.929454] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.929454] env[61985]: raise e [ 669.929454] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.929454] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 669.929454] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.929454] env[61985]: created_port_ids = self._update_ports_for_instance( [ 669.929454] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.929454] env[61985]: with excutils.save_and_reraise_exception(): [ 669.929454] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.929454] env[61985]: self.force_reraise() [ 669.929454] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.929454] env[61985]: raise self.value [ 669.929454] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.929454] env[61985]: updated_port = self._update_port( [ 669.929454] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.929454] env[61985]: _ensure_no_port_binding_failure(port) [ 669.929454] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.929454] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.930189] env[61985]: nova.exception.PortBindingFailed: Binding failed for port a1fe1a7b-3745-42c6-9f72-0afc45077ddf, please check neutron logs for more information. [ 669.930189] env[61985]: Removing descriptor: 17 [ 669.930189] env[61985]: ERROR nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a1fe1a7b-3745-42c6-9f72-0afc45077ddf, please check neutron logs for more information. [ 669.930189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Traceback (most recent call last): [ 669.930189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 669.930189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] yield resources [ 669.930189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 669.930189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] self.driver.spawn(context, instance, image_meta, [ 669.930189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.930189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.930189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.930189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] vm_ref = self.build_virtual_machine(instance, [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] for vif in network_info: [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] return self._sync_wrapper(fn, *args, **kwargs) [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] self.wait() [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] self[:] = self._gt.wait() [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] return self._exit_event.wait() [ 669.930482] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] result = hub.switch() [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] return self.greenlet.switch() [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] result = function(*args, **kwargs) [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] return func(*args, **kwargs) [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] raise e [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] nwinfo = self.network_api.allocate_for_instance( [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.930765] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] created_port_ids = self._update_ports_for_instance( [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] with excutils.save_and_reraise_exception(): [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] self.force_reraise() [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] raise self.value [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] updated_port = self._update_port( [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] _ensure_no_port_binding_failure(port) [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.931039] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] raise exception.PortBindingFailed(port_id=port['id']) [ 669.931296] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] nova.exception.PortBindingFailed: Binding failed for port a1fe1a7b-3745-42c6-9f72-0afc45077ddf, please check neutron logs for more information. [ 669.931296] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] [ 669.931296] env[61985]: INFO nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Terminating instance [ 669.931919] env[61985]: DEBUG nova.network.neutron [-] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.933699] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "refresh_cache-c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.933699] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "refresh_cache-c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.933699] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.374027] env[61985]: DEBUG nova.compute.utils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 670.379024] env[61985]: DEBUG nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 670.379024] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 670.421062] env[61985]: DEBUG nova.policy [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2dba42e753b4641862b3a03ce563892', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff085eb00ec4486b923bcdfd67406b14', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 670.435388] env[61985]: INFO nova.compute.manager [-] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Took 1.03 seconds to deallocate network for instance. [ 670.438330] env[61985]: DEBUG nova.compute.claims [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 670.438869] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.462045] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.607986] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.831508] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Successfully created port: 2000fbab-b01f-4a68-bef7-1d5f32773dc1 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 670.879453] env[61985]: DEBUG nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 671.111820] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "refresh_cache-c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.112867] env[61985]: DEBUG nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 671.113103] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 671.113605] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-470b68d9-5507-403f-bd9e-d6fdee68f563 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.128846] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99672b4-6767-4364-bed2-6e5195768ee6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.147381] env[61985]: DEBUG nova.compute.manager [req-c0b0a4d2-cad5-459e-a598-82f8c3b4ee87 req-7f56aed9-ca7f-444c-be1e-eec70ebd0a46 service nova] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Received event network-changed-a1fe1a7b-3745-42c6-9f72-0afc45077ddf {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 671.147622] env[61985]: DEBUG nova.compute.manager [req-c0b0a4d2-cad5-459e-a598-82f8c3b4ee87 req-7f56aed9-ca7f-444c-be1e-eec70ebd0a46 service nova] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Refreshing instance network info cache due to event network-changed-a1fe1a7b-3745-42c6-9f72-0afc45077ddf. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 671.147830] env[61985]: DEBUG oslo_concurrency.lockutils [req-c0b0a4d2-cad5-459e-a598-82f8c3b4ee87 req-7f56aed9-ca7f-444c-be1e-eec70ebd0a46 service nova] Acquiring lock "refresh_cache-c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.147965] env[61985]: DEBUG oslo_concurrency.lockutils [req-c0b0a4d2-cad5-459e-a598-82f8c3b4ee87 req-7f56aed9-ca7f-444c-be1e-eec70ebd0a46 service nova] Acquired lock "refresh_cache-c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.148142] env[61985]: DEBUG nova.network.neutron [req-c0b0a4d2-cad5-459e-a598-82f8c3b4ee87 req-7f56aed9-ca7f-444c-be1e-eec70ebd0a46 service nova] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Refreshing network info cache for port a1fe1a7b-3745-42c6-9f72-0afc45077ddf {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 671.160946] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124 could not be found. [ 671.161086] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 671.161293] env[61985]: INFO nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Took 0.05 seconds to destroy the instance on the hypervisor. [ 671.161584] env[61985]: DEBUG oslo.service.loopingcall [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.164851] env[61985]: DEBUG nova.compute.manager [-] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.164959] env[61985]: DEBUG nova.network.neutron [-] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 671.199095] env[61985]: DEBUG nova.network.neutron [-] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.429511] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b03e6e-0bdd-4f4b-a188-8ad60fcce0ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.439269] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d877471-c323-4ca7-86c0-5e9d45e2a5f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.474389] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9ae850-a8b6-4d4c-bf72-2ba6508e248b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.483810] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0200810-6b9d-4ae3-a741-d091b0643bbb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.500981] env[61985]: DEBUG nova.compute.provider_tree [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.684224] env[61985]: DEBUG nova.network.neutron [req-c0b0a4d2-cad5-459e-a598-82f8c3b4ee87 req-7f56aed9-ca7f-444c-be1e-eec70ebd0a46 service nova] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.703024] env[61985]: DEBUG nova.network.neutron [-] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.876706] env[61985]: DEBUG nova.network.neutron [req-c0b0a4d2-cad5-459e-a598-82f8c3b4ee87 req-7f56aed9-ca7f-444c-be1e-eec70ebd0a46 service nova] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.891492] env[61985]: DEBUG nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 671.924846] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 671.925516] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 671.925516] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 671.925516] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 671.925713] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 671.925758] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 671.925956] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 671.926122] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 671.926281] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 671.926673] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 671.926673] env[61985]: DEBUG nova.virt.hardware [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.928083] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ca3a94-44dd-472e-bbbb-b0e6cadc6375 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.937019] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3561e331-c088-4a51-b638-acdf498bbba9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.005703] env[61985]: DEBUG nova.scheduler.client.report [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.207018] env[61985]: INFO nova.compute.manager [-] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Took 1.04 seconds to deallocate network for instance. [ 672.214130] env[61985]: DEBUG nova.compute.claims [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 672.214130] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.380831] env[61985]: DEBUG oslo_concurrency.lockutils [req-c0b0a4d2-cad5-459e-a598-82f8c3b4ee87 req-7f56aed9-ca7f-444c-be1e-eec70ebd0a46 service nova] Releasing lock "refresh_cache-c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.380831] env[61985]: DEBUG nova.compute.manager [req-c0b0a4d2-cad5-459e-a598-82f8c3b4ee87 req-7f56aed9-ca7f-444c-be1e-eec70ebd0a46 service nova] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Received event network-vif-deleted-a1fe1a7b-3745-42c6-9f72-0afc45077ddf {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 672.511919] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.652s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.512359] env[61985]: DEBUG nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 672.515183] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.384s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.516558] env[61985]: INFO nova.compute.claims [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 672.608854] env[61985]: ERROR nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. [ 672.608854] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 672.608854] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.608854] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 672.608854] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 672.608854] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 672.608854] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 672.608854] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 672.608854] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.608854] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 672.608854] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.608854] env[61985]: ERROR nova.compute.manager raise self.value [ 672.608854] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 672.608854] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 672.608854] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.608854] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 672.609323] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.609323] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 672.609323] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. [ 672.609323] env[61985]: ERROR nova.compute.manager [ 672.609323] env[61985]: Traceback (most recent call last): [ 672.609323] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 672.609323] env[61985]: listener.cb(fileno) [ 672.609323] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.609323] env[61985]: result = function(*args, **kwargs) [ 672.609323] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.609323] env[61985]: return func(*args, **kwargs) [ 672.609323] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.609323] env[61985]: raise e [ 672.609323] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.609323] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 672.609323] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 672.609323] env[61985]: created_port_ids = self._update_ports_for_instance( [ 672.609323] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 672.609323] env[61985]: with excutils.save_and_reraise_exception(): [ 672.609323] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.609323] env[61985]: self.force_reraise() [ 672.609323] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.609323] env[61985]: raise self.value [ 672.609323] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 672.609323] env[61985]: updated_port = self._update_port( [ 672.609323] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.609323] env[61985]: _ensure_no_port_binding_failure(port) [ 672.609323] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.609323] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 672.609976] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. [ 672.609976] env[61985]: Removing descriptor: 17 [ 672.609976] env[61985]: ERROR nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. [ 672.609976] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Traceback (most recent call last): [ 672.609976] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 672.609976] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] yield resources [ 672.609976] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 672.609976] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] self.driver.spawn(context, instance, image_meta, [ 672.609976] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 672.609976] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.609976] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.609976] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] vm_ref = self.build_virtual_machine(instance, [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] for vif in network_info: [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] return self._sync_wrapper(fn, *args, **kwargs) [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] self.wait() [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] self[:] = self._gt.wait() [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] return self._exit_event.wait() [ 672.610261] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] result = hub.switch() [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] return self.greenlet.switch() [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] result = function(*args, **kwargs) [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] return func(*args, **kwargs) [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] raise e [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] nwinfo = self.network_api.allocate_for_instance( [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 672.610593] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] created_port_ids = self._update_ports_for_instance( [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] with excutils.save_and_reraise_exception(): [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] self.force_reraise() [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] raise self.value [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] updated_port = self._update_port( [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] _ensure_no_port_binding_failure(port) [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.610970] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] raise exception.PortBindingFailed(port_id=port['id']) [ 672.611283] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] nova.exception.PortBindingFailed: Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. [ 672.611283] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] [ 672.611283] env[61985]: INFO nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Terminating instance [ 672.612377] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "refresh_cache-955181b1-eb46-424c-8c79-055638dc9dac" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.612586] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquired lock "refresh_cache-955181b1-eb46-424c-8c79-055638dc9dac" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.612788] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 673.022117] env[61985]: DEBUG nova.compute.utils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.026306] env[61985]: DEBUG nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.026543] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 673.078436] env[61985]: DEBUG nova.policy [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a78dbc667d24c819e6447740fc2d8e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2dbceebf07c43ac90ed80f8a09294b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.142789] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.159066] env[61985]: DEBUG nova.compute.manager [req-aa52c633-937c-4fb7-ba84-d5fec7613bff req-06f46a99-6809-4092-8377-054293538102 service nova] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Received event network-changed-2000fbab-b01f-4a68-bef7-1d5f32773dc1 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 673.159273] env[61985]: DEBUG nova.compute.manager [req-aa52c633-937c-4fb7-ba84-d5fec7613bff req-06f46a99-6809-4092-8377-054293538102 service nova] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Refreshing instance network info cache due to event network-changed-2000fbab-b01f-4a68-bef7-1d5f32773dc1. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 673.159637] env[61985]: DEBUG oslo_concurrency.lockutils [req-aa52c633-937c-4fb7-ba84-d5fec7613bff req-06f46a99-6809-4092-8377-054293538102 service nova] Acquiring lock "refresh_cache-955181b1-eb46-424c-8c79-055638dc9dac" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.345131] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.530395] env[61985]: DEBUG nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 673.623650] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Successfully created port: f5e4273b-4de4-4a93-a78c-4615fa130904 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.850789] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Releasing lock "refresh_cache-955181b1-eb46-424c-8c79-055638dc9dac" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.850789] env[61985]: DEBUG nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 673.853388] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 673.853388] env[61985]: DEBUG oslo_concurrency.lockutils [req-aa52c633-937c-4fb7-ba84-d5fec7613bff req-06f46a99-6809-4092-8377-054293538102 service nova] Acquired lock "refresh_cache-955181b1-eb46-424c-8c79-055638dc9dac" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.853388] env[61985]: DEBUG nova.network.neutron [req-aa52c633-937c-4fb7-ba84-d5fec7613bff req-06f46a99-6809-4092-8377-054293538102 service nova] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Refreshing network info cache for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 673.853388] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2f54a2e9-bdc0-4d41-ad2d-e49aac7ad1cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.869734] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753b9c02-bb6e-48ce-8162-b40f4e2e6337 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.892236] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 955181b1-eb46-424c-8c79-055638dc9dac could not be found. [ 673.892491] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 673.892654] env[61985]: INFO nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Took 0.04 seconds to destroy the instance on the hypervisor. [ 673.892885] env[61985]: DEBUG oslo.service.loopingcall [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.895356] env[61985]: DEBUG nova.compute.manager [-] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.895463] env[61985]: DEBUG nova.network.neutron [-] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 674.017146] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6286effd-fbce-4339-9548-aaaee292f97c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.023942] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab705ff-5c69-4c03-a626-4d953ca24588 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.061284] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4aecc32-7dc5-40e9-b18e-9cc5c43c26ac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.073020] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660726a4-d94d-4ada-a21d-e281bcaee6f0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.084369] env[61985]: DEBUG nova.compute.provider_tree [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.094735] env[61985]: DEBUG nova.network.neutron [-] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.390011] env[61985]: DEBUG nova.network.neutron [req-aa52c633-937c-4fb7-ba84-d5fec7613bff req-06f46a99-6809-4092-8377-054293538102 service nova] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.565667] env[61985]: DEBUG nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 674.568614] env[61985]: DEBUG nova.network.neutron [req-aa52c633-937c-4fb7-ba84-d5fec7613bff req-06f46a99-6809-4092-8377-054293538102 service nova] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.587761] env[61985]: DEBUG nova.scheduler.client.report [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.598735] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 674.598989] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 674.599163] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.599348] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 674.599487] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.599681] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 674.599900] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 674.600068] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 674.600233] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 674.600389] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 674.600557] env[61985]: DEBUG nova.virt.hardware [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.600842] env[61985]: DEBUG nova.network.neutron [-] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.605719] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c9cac3-4726-4328-9fa9-652e3c6fd2c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.614934] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79829129-55f6-4a9a-8d4a-662476a6f599 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.070942] env[61985]: DEBUG oslo_concurrency.lockutils [req-aa52c633-937c-4fb7-ba84-d5fec7613bff req-06f46a99-6809-4092-8377-054293538102 service nova] Releasing lock "refresh_cache-955181b1-eb46-424c-8c79-055638dc9dac" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.071228] env[61985]: DEBUG nova.compute.manager [req-aa52c633-937c-4fb7-ba84-d5fec7613bff req-06f46a99-6809-4092-8377-054293538102 service nova] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Received event network-vif-deleted-2000fbab-b01f-4a68-bef7-1d5f32773dc1 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 675.095441] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.580s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.096137] env[61985]: DEBUG nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 675.098832] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.927s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.100251] env[61985]: INFO nova.compute.claims [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.110124] env[61985]: INFO nova.compute.manager [-] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Took 1.21 seconds to deallocate network for instance. [ 675.112364] env[61985]: ERROR nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f5e4273b-4de4-4a93-a78c-4615fa130904, please check neutron logs for more information. [ 675.112364] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 675.112364] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.112364] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 675.112364] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.112364] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 675.112364] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.112364] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 675.112364] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.112364] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 675.112364] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.112364] env[61985]: ERROR nova.compute.manager raise self.value [ 675.112364] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.112364] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 675.112364] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.112364] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 675.112828] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.112828] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 675.112828] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f5e4273b-4de4-4a93-a78c-4615fa130904, please check neutron logs for more information. [ 675.112828] env[61985]: ERROR nova.compute.manager [ 675.112828] env[61985]: Traceback (most recent call last): [ 675.112828] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 675.112828] env[61985]: listener.cb(fileno) [ 675.112828] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.112828] env[61985]: result = function(*args, **kwargs) [ 675.112828] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.112828] env[61985]: return func(*args, **kwargs) [ 675.112828] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.112828] env[61985]: raise e [ 675.112828] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.112828] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 675.112828] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.112828] env[61985]: created_port_ids = self._update_ports_for_instance( [ 675.112828] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.112828] env[61985]: with excutils.save_and_reraise_exception(): [ 675.112828] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.112828] env[61985]: self.force_reraise() [ 675.112828] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.112828] env[61985]: raise self.value [ 675.112828] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.112828] env[61985]: updated_port = self._update_port( [ 675.112828] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.112828] env[61985]: _ensure_no_port_binding_failure(port) [ 675.112828] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.112828] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 675.113557] env[61985]: nova.exception.PortBindingFailed: Binding failed for port f5e4273b-4de4-4a93-a78c-4615fa130904, please check neutron logs for more information. [ 675.113557] env[61985]: Removing descriptor: 17 [ 675.113557] env[61985]: ERROR nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f5e4273b-4de4-4a93-a78c-4615fa130904, please check neutron logs for more information. [ 675.113557] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Traceback (most recent call last): [ 675.113557] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 675.113557] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] yield resources [ 675.113557] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 675.113557] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] self.driver.spawn(context, instance, image_meta, [ 675.113557] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.113557] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.113557] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.113557] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] vm_ref = self.build_virtual_machine(instance, [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] for vif in network_info: [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] return self._sync_wrapper(fn, *args, **kwargs) [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] self.wait() [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] self[:] = self._gt.wait() [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] return self._exit_event.wait() [ 675.113943] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] result = hub.switch() [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] return self.greenlet.switch() [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] result = function(*args, **kwargs) [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] return func(*args, **kwargs) [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] raise e [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] nwinfo = self.network_api.allocate_for_instance( [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.114271] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] created_port_ids = self._update_ports_for_instance( [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] with excutils.save_and_reraise_exception(): [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] self.force_reraise() [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] raise self.value [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] updated_port = self._update_port( [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] _ensure_no_port_binding_failure(port) [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.114667] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] raise exception.PortBindingFailed(port_id=port['id']) [ 675.114986] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] nova.exception.PortBindingFailed: Binding failed for port f5e4273b-4de4-4a93-a78c-4615fa130904, please check neutron logs for more information. [ 675.114986] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] [ 675.114986] env[61985]: INFO nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Terminating instance [ 675.115341] env[61985]: DEBUG nova.compute.claims [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 675.115520] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.125986] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-f86a5054-f30e-4868-9c50-2fe0d384b0b0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.126251] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-f86a5054-f30e-4868-9c50-2fe0d384b0b0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.126424] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 675.185461] env[61985]: DEBUG nova.compute.manager [req-15ad017c-94d0-4ef0-8b5c-a43ef859c46f req-7ba75341-0fa1-4e61-85c5-d1603584dad3 service nova] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Received event network-changed-f5e4273b-4de4-4a93-a78c-4615fa130904 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 675.185658] env[61985]: DEBUG nova.compute.manager [req-15ad017c-94d0-4ef0-8b5c-a43ef859c46f req-7ba75341-0fa1-4e61-85c5-d1603584dad3 service nova] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Refreshing instance network info cache due to event network-changed-f5e4273b-4de4-4a93-a78c-4615fa130904. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 675.185852] env[61985]: DEBUG oslo_concurrency.lockutils [req-15ad017c-94d0-4ef0-8b5c-a43ef859c46f req-7ba75341-0fa1-4e61-85c5-d1603584dad3 service nova] Acquiring lock "refresh_cache-f86a5054-f30e-4868-9c50-2fe0d384b0b0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.604791] env[61985]: DEBUG nova.compute.utils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.608250] env[61985]: DEBUG nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.608422] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 675.654556] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.719321] env[61985]: DEBUG nova.policy [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '30de2f5bbdad48a3b2c90866025d6a65', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5af338aacaea4ec6b489957b327f41c8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 675.811911] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.109357] env[61985]: DEBUG nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 676.187467] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Successfully created port: b2ee6e34-2ef1-4414-b847-2028086a3da9 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.315313] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-f86a5054-f30e-4868-9c50-2fe0d384b0b0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.315732] env[61985]: DEBUG nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 676.315926] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 676.316488] env[61985]: DEBUG oslo_concurrency.lockutils [req-15ad017c-94d0-4ef0-8b5c-a43ef859c46f req-7ba75341-0fa1-4e61-85c5-d1603584dad3 service nova] Acquired lock "refresh_cache-f86a5054-f30e-4868-9c50-2fe0d384b0b0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.316659] env[61985]: DEBUG nova.network.neutron [req-15ad017c-94d0-4ef0-8b5c-a43ef859c46f req-7ba75341-0fa1-4e61-85c5-d1603584dad3 service nova] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Refreshing network info cache for port f5e4273b-4de4-4a93-a78c-4615fa130904 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 676.317706] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8fe960e6-15d7-49c8-8c3d-cc840ac9e5d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.326215] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a7359e-6ba0-4f41-a0b6-4414a0490b13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.359396] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f86a5054-f30e-4868-9c50-2fe0d384b0b0 could not be found. [ 676.359641] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 676.363015] env[61985]: INFO nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 676.363015] env[61985]: DEBUG oslo.service.loopingcall [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.363015] env[61985]: DEBUG nova.compute.manager [-] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.363015] env[61985]: DEBUG nova.network.neutron [-] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 676.388064] env[61985]: DEBUG nova.network.neutron [-] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.037059] env[61985]: DEBUG nova.network.neutron [-] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.057269] env[61985]: DEBUG nova.network.neutron [req-15ad017c-94d0-4ef0-8b5c-a43ef859c46f req-7ba75341-0fa1-4e61-85c5-d1603584dad3 service nova] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.156099] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b444ca9e-e26e-4e3d-a897-ff00064dddaf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.163487] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa052df-731f-49f9-9e29-9ddb6e225656 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.192604] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66034cba-d857-48f7-a064-ad0cd4670588 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.199513] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55735703-121e-4327-9f7e-39565017f182 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.203630] env[61985]: DEBUG nova.network.neutron [req-15ad017c-94d0-4ef0-8b5c-a43ef859c46f req-7ba75341-0fa1-4e61-85c5-d1603584dad3 service nova] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.214166] env[61985]: DEBUG nova.compute.provider_tree [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.322457] env[61985]: DEBUG nova.compute.manager [req-bb821561-aeaa-4f4f-b974-5aa4d5af2ec0 req-d80e170d-1702-44f9-95d6-4822e0382ffb service nova] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Received event network-changed-b2ee6e34-2ef1-4414-b847-2028086a3da9 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 677.322657] env[61985]: DEBUG nova.compute.manager [req-bb821561-aeaa-4f4f-b974-5aa4d5af2ec0 req-d80e170d-1702-44f9-95d6-4822e0382ffb service nova] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Refreshing instance network info cache due to event network-changed-b2ee6e34-2ef1-4414-b847-2028086a3da9. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 677.322892] env[61985]: DEBUG oslo_concurrency.lockutils [req-bb821561-aeaa-4f4f-b974-5aa4d5af2ec0 req-d80e170d-1702-44f9-95d6-4822e0382ffb service nova] Acquiring lock "refresh_cache-033766e8-92c0-498d-b3e8-37de5b46c20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.323016] env[61985]: DEBUG oslo_concurrency.lockutils [req-bb821561-aeaa-4f4f-b974-5aa4d5af2ec0 req-d80e170d-1702-44f9-95d6-4822e0382ffb service nova] Acquired lock "refresh_cache-033766e8-92c0-498d-b3e8-37de5b46c20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.323188] env[61985]: DEBUG nova.network.neutron [req-bb821561-aeaa-4f4f-b974-5aa4d5af2ec0 req-d80e170d-1702-44f9-95d6-4822e0382ffb service nova] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Refreshing network info cache for port b2ee6e34-2ef1-4414-b847-2028086a3da9 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 677.525040] env[61985]: ERROR nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b2ee6e34-2ef1-4414-b847-2028086a3da9, please check neutron logs for more information. [ 677.525040] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 677.525040] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.525040] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 677.525040] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.525040] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 677.525040] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.525040] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 677.525040] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.525040] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 677.525040] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.525040] env[61985]: ERROR nova.compute.manager raise self.value [ 677.525040] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.525040] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 677.525040] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.525040] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 677.525498] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.525498] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 677.525498] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b2ee6e34-2ef1-4414-b847-2028086a3da9, please check neutron logs for more information. [ 677.525498] env[61985]: ERROR nova.compute.manager [ 677.525498] env[61985]: Traceback (most recent call last): [ 677.525498] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 677.525498] env[61985]: listener.cb(fileno) [ 677.525498] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.525498] env[61985]: result = function(*args, **kwargs) [ 677.525498] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.525498] env[61985]: return func(*args, **kwargs) [ 677.525498] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.525498] env[61985]: raise e [ 677.525498] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.525498] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 677.525498] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.525498] env[61985]: created_port_ids = self._update_ports_for_instance( [ 677.525498] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.525498] env[61985]: with excutils.save_and_reraise_exception(): [ 677.525498] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.525498] env[61985]: self.force_reraise() [ 677.525498] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.525498] env[61985]: raise self.value [ 677.525498] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.525498] env[61985]: updated_port = self._update_port( [ 677.525498] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.525498] env[61985]: _ensure_no_port_binding_failure(port) [ 677.525498] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.525498] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 677.526256] env[61985]: nova.exception.PortBindingFailed: Binding failed for port b2ee6e34-2ef1-4414-b847-2028086a3da9, please check neutron logs for more information. [ 677.526256] env[61985]: Removing descriptor: 17 [ 677.539680] env[61985]: DEBUG nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 677.542116] env[61985]: INFO nova.compute.manager [-] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Took 1.18 seconds to deallocate network for instance. [ 677.545653] env[61985]: DEBUG nova.compute.claims [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 677.545837] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.567114] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.567114] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.567629] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.567858] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.569259] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.569431] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.569647] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.569812] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.570270] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.570270] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.570357] env[61985]: DEBUG nova.virt.hardware [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.571137] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76d083a-7295-457a-9009-09a1e17515c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.578982] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb7860c-ee09-48a0-b64d-cfd18b674f86 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.592035] env[61985]: ERROR nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b2ee6e34-2ef1-4414-b847-2028086a3da9, please check neutron logs for more information. [ 677.592035] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Traceback (most recent call last): [ 677.592035] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 677.592035] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] yield resources [ 677.592035] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 677.592035] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] self.driver.spawn(context, instance, image_meta, [ 677.592035] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.592035] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.592035] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.592035] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] vm_ref = self.build_virtual_machine(instance, [ 677.592035] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] for vif in network_info: [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] return self._sync_wrapper(fn, *args, **kwargs) [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] self.wait() [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] self[:] = self._gt.wait() [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] return self._exit_event.wait() [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.592357] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] current.throw(*self._exc) [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] result = function(*args, **kwargs) [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] return func(*args, **kwargs) [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] raise e [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] nwinfo = self.network_api.allocate_for_instance( [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] created_port_ids = self._update_ports_for_instance( [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] with excutils.save_and_reraise_exception(): [ 677.592674] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] self.force_reraise() [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] raise self.value [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] updated_port = self._update_port( [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] _ensure_no_port_binding_failure(port) [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] raise exception.PortBindingFailed(port_id=port['id']) [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] nova.exception.PortBindingFailed: Binding failed for port b2ee6e34-2ef1-4414-b847-2028086a3da9, please check neutron logs for more information. [ 677.593029] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] [ 677.593029] env[61985]: INFO nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Terminating instance [ 677.594187] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Acquiring lock "refresh_cache-033766e8-92c0-498d-b3e8-37de5b46c20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.706036] env[61985]: DEBUG oslo_concurrency.lockutils [req-15ad017c-94d0-4ef0-8b5c-a43ef859c46f req-7ba75341-0fa1-4e61-85c5-d1603584dad3 service nova] Releasing lock "refresh_cache-f86a5054-f30e-4868-9c50-2fe0d384b0b0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.706353] env[61985]: DEBUG nova.compute.manager [req-15ad017c-94d0-4ef0-8b5c-a43ef859c46f req-7ba75341-0fa1-4e61-85c5-d1603584dad3 service nova] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Received event network-vif-deleted-f5e4273b-4de4-4a93-a78c-4615fa130904 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 677.716488] env[61985]: DEBUG nova.scheduler.client.report [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.851468] env[61985]: DEBUG nova.network.neutron [req-bb821561-aeaa-4f4f-b974-5aa4d5af2ec0 req-d80e170d-1702-44f9-95d6-4822e0382ffb service nova] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.927684] env[61985]: DEBUG nova.network.neutron [req-bb821561-aeaa-4f4f-b974-5aa4d5af2ec0 req-d80e170d-1702-44f9-95d6-4822e0382ffb service nova] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.221944] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.123s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.222531] env[61985]: DEBUG nova.compute.manager [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 678.225513] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.835s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.430383] env[61985]: DEBUG oslo_concurrency.lockutils [req-bb821561-aeaa-4f4f-b974-5aa4d5af2ec0 req-d80e170d-1702-44f9-95d6-4822e0382ffb service nova] Releasing lock "refresh_cache-033766e8-92c0-498d-b3e8-37de5b46c20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.430879] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Acquired lock "refresh_cache-033766e8-92c0-498d-b3e8-37de5b46c20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.431082] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.730223] env[61985]: DEBUG nova.compute.utils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.736647] env[61985]: DEBUG nova.compute.manager [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Not allocating networking since 'none' was specified. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 678.951191] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.037541] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.107997] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0cb38d9-d4e9-42f5-beac-1f040e50a456 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.115599] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8976d13e-a9ac-499c-91bb-85ace5068d76 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.147138] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654cfdc8-e938-40df-bedd-577d73494c85 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.154246] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1432fa7e-048c-4001-95aa-9f92383dde25 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.167383] env[61985]: DEBUG nova.compute.provider_tree [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.233362] env[61985]: DEBUG nova.compute.manager [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 679.348873] env[61985]: DEBUG nova.compute.manager [req-18c424d7-9b41-4c5a-ac27-ecaf80e4898f req-e76dfb66-4010-4142-a20c-760f344bfbee service nova] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Received event network-vif-deleted-b2ee6e34-2ef1-4414-b847-2028086a3da9 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 679.540028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Releasing lock "refresh_cache-033766e8-92c0-498d-b3e8-37de5b46c20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.540240] env[61985]: DEBUG nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 679.540437] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 679.540729] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bedfafd5-d1de-455a-9f89-bc407beaf464 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.549469] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918bb06d-85f1-4caa-bcd0-d9f4ed753dcf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.570349] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 033766e8-92c0-498d-b3e8-37de5b46c20b could not be found. [ 679.571263] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 679.571457] env[61985]: INFO nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 679.571699] env[61985]: DEBUG oslo.service.loopingcall [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.572369] env[61985]: DEBUG nova.compute.manager [-] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.572369] env[61985]: DEBUG nova.network.neutron [-] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 679.586998] env[61985]: DEBUG nova.network.neutron [-] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.673152] env[61985]: DEBUG nova.scheduler.client.report [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.089917] env[61985]: DEBUG nova.network.neutron [-] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.178792] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.953s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.178954] env[61985]: ERROR nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 36457600-3fc2-43e3-a354-7f28da7c3e44, please check neutron logs for more information. [ 680.178954] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Traceback (most recent call last): [ 680.178954] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 680.178954] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] self.driver.spawn(context, instance, image_meta, [ 680.178954] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.178954] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.178954] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.178954] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] vm_ref = self.build_virtual_machine(instance, [ 680.178954] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.178954] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.178954] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] for vif in network_info: [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] return self._sync_wrapper(fn, *args, **kwargs) [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] self.wait() [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] self[:] = self._gt.wait() [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] return self._exit_event.wait() [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] result = hub.switch() [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.179270] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] return self.greenlet.switch() [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] result = function(*args, **kwargs) [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] return func(*args, **kwargs) [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] raise e [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] nwinfo = self.network_api.allocate_for_instance( [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] created_port_ids = self._update_ports_for_instance( [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] with excutils.save_and_reraise_exception(): [ 680.179656] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] self.force_reraise() [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] raise self.value [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] updated_port = self._update_port( [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] _ensure_no_port_binding_failure(port) [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] raise exception.PortBindingFailed(port_id=port['id']) [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] nova.exception.PortBindingFailed: Binding failed for port 36457600-3fc2-43e3-a354-7f28da7c3e44, please check neutron logs for more information. [ 680.180129] env[61985]: ERROR nova.compute.manager [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] [ 680.180488] env[61985]: DEBUG nova.compute.utils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Binding failed for port 36457600-3fc2-43e3-a354-7f28da7c3e44, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.181108] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.580s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.181333] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.181426] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 680.181697] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.833s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.185030] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484acb53-3096-48f5-b063-5dcfd3f1bfb1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.189375] env[61985]: DEBUG nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Build of instance 9ac9f4b8-fcef-458a-913c-72cfc3322b6e was re-scheduled: Binding failed for port 36457600-3fc2-43e3-a354-7f28da7c3e44, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 680.189824] env[61985]: DEBUG nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 680.190062] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Acquiring lock "refresh_cache-9ac9f4b8-fcef-458a-913c-72cfc3322b6e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.190211] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Acquired lock "refresh_cache-9ac9f4b8-fcef-458a-913c-72cfc3322b6e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.190400] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 680.195922] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74fa088-ba43-47d6-abc9-49fea53d553f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.210791] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ae968d-7b56-4a15-a8f7-3427832f67c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.217236] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916ce8d8-7fd2-48ff-8a89-6c3bb8c951f9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.246334] env[61985]: DEBUG nova.compute.manager [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 680.248396] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181481MB free_disk=44GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 680.248476] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.266561] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 680.266561] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 680.266561] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 680.266754] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 680.266867] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 680.267099] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 680.267327] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 680.267514] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 680.267690] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 680.267855] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 680.268034] env[61985]: DEBUG nova.virt.hardware [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 680.268792] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26ff3e6-394a-4779-b9f8-f9041183d44c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.275989] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c467e1-aea7-499d-b8e0-1499742e8835 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.288629] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 680.295723] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Creating folder: Project (2535f3b548444ee3b6d558e2f2cfe207). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 680.295723] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b35ffc2-b2ab-49a5-8bd2-ca35cd98987c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.304624] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Created folder: Project (2535f3b548444ee3b6d558e2f2cfe207) in parent group-v211285. [ 680.304815] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Creating folder: Instances. Parent ref: group-v211293. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 680.305039] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6d85125-1612-4d46-a009-5e216c4258ae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.312892] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Created folder: Instances in parent group-v211293. [ 680.313147] env[61985]: DEBUG oslo.service.loopingcall [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 680.313347] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 680.313555] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7de393dd-cc75-42dd-9e92-06b82963212d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.328828] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 680.328828] env[61985]: value = "task-935626" [ 680.328828] env[61985]: _type = "Task" [ 680.328828] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.335504] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935626, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.595081] env[61985]: INFO nova.compute.manager [-] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Took 1.02 seconds to deallocate network for instance. [ 680.597950] env[61985]: DEBUG nova.compute.claims [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 680.598181] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.713851] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.835016] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.840435] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935626, 'name': CreateVM_Task, 'duration_secs': 0.257125} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.842994] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 680.844280] env[61985]: DEBUG oslo_vmware.service [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2c2a0b-7dc3-4efc-92e9-a30924ff4249 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.850791] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.850956] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.851342] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 680.851798] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43a1818a-695e-41d1-bfdf-3498bcc1329a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.860372] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 680.860372] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522d3977-9a26-8b70-dcda-3c400c8c572c" [ 680.860372] env[61985]: _type = "Task" [ 680.860372] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.868195] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522d3977-9a26-8b70-dcda-3c400c8c572c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.074751] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06cf098-c803-4e93-a9c6-d2b9220ab97e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.082354] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7448c3-bae9-459d-967b-73c9d8e219c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.112193] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc0f036-604b-4722-8556-a6b1c4c47e61 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.119423] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95749b88-f21c-4894-86b4-0745630ae48f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.132202] env[61985]: DEBUG nova.compute.provider_tree [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.344022] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Releasing lock "refresh_cache-9ac9f4b8-fcef-458a-913c-72cfc3322b6e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.344340] env[61985]: DEBUG nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 681.344536] env[61985]: DEBUG nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.344699] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.359803] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.370751] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.370985] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 681.371228] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.371371] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.371540] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 681.372011] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dcf803be-8409-4301-b995-14f66f62d930 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.389968] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 681.389968] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 681.390714] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb082302-f889-460d-a535-db7f9473242e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.396823] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a8d9a1c-6d41-48ae-ba68-f7fe3aefa822 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.403240] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 681.403240] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52bd514f-6c5a-20ef-0db1-7b185683b3ee" [ 681.403240] env[61985]: _type = "Task" [ 681.403240] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.410522] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52bd514f-6c5a-20ef-0db1-7b185683b3ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.635095] env[61985]: DEBUG nova.scheduler.client.report [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.865907] env[61985]: DEBUG nova.network.neutron [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.914450] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Preparing fetch location {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 681.914724] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Creating directory with path [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 681.914958] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84f12099-e36d-4134-9345-672f2f9d7ab6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.939790] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Created directory with path [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 681.940043] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Fetch image to [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 681.940227] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Downloading image file data fe6d817b-6194-440f-988a-f4a94c580922 to [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk on the data store datastore1 {{(pid=61985) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 681.941301] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feeff51d-ac96-4f27-814e-b77f1f2793a0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.947919] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725e2dff-17d3-4f03-92d0-50d29c169ddf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.957039] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8306fb-b4c8-488d-a72a-b3fa0d0623a9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.987670] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4335053-9605-494d-afd7-e57be1902cac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.993511] env[61985]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-59b4542d-6413-441a-b5f6-420c1b70706c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.015955] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Downloading image file data fe6d817b-6194-440f-988a-f4a94c580922 to the data store datastore1 {{(pid=61985) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 682.060867] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61985) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 682.140076] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.140606] env[61985]: ERROR nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8d895f96-2327-47ba-a747-e6b97832e9fc, please check neutron logs for more information. [ 682.140606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Traceback (most recent call last): [ 682.140606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 682.140606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] self.driver.spawn(context, instance, image_meta, [ 682.140606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 682.140606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.140606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.140606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] vm_ref = self.build_virtual_machine(instance, [ 682.140606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.140606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.140606] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] for vif in network_info: [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] return self._sync_wrapper(fn, *args, **kwargs) [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] self.wait() [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] self[:] = self._gt.wait() [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] return self._exit_event.wait() [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] result = hub.switch() [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 682.140929] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] return self.greenlet.switch() [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] result = function(*args, **kwargs) [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] return func(*args, **kwargs) [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] raise e [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] nwinfo = self.network_api.allocate_for_instance( [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] created_port_ids = self._update_ports_for_instance( [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] with excutils.save_and_reraise_exception(): [ 682.141292] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] self.force_reraise() [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] raise self.value [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] updated_port = self._update_port( [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] _ensure_no_port_binding_failure(port) [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] raise exception.PortBindingFailed(port_id=port['id']) [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] nova.exception.PortBindingFailed: Binding failed for port 8d895f96-2327-47ba-a747-e6b97832e9fc, please check neutron logs for more information. [ 682.141660] env[61985]: ERROR nova.compute.manager [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] [ 682.142025] env[61985]: DEBUG nova.compute.utils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Binding failed for port 8d895f96-2327-47ba-a747-e6b97832e9fc, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.142501] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.531s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.143985] env[61985]: INFO nova.compute.claims [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.147351] env[61985]: DEBUG nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Build of instance 7f32151e-7be0-40e2-b210-72765b7551bc was re-scheduled: Binding failed for port 8d895f96-2327-47ba-a747-e6b97832e9fc, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 682.147841] env[61985]: DEBUG nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 682.148092] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquiring lock "refresh_cache-7f32151e-7be0-40e2-b210-72765b7551bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.148243] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Acquired lock "refresh_cache-7f32151e-7be0-40e2-b210-72765b7551bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.148399] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 682.368473] env[61985]: INFO nova.compute.manager [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] [instance: 9ac9f4b8-fcef-458a-913c-72cfc3322b6e] Took 1.02 seconds to deallocate network for instance. [ 682.675738] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.711195] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Completed reading data from the image iterator. {{(pid=61985) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 682.711455] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 682.746789] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.829089] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Downloaded image file data fe6d817b-6194-440f-988a-f4a94c580922 to vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk on the data store datastore1 {{(pid=61985) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 682.831428] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Caching image {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 682.831698] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Copying Virtual Disk [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk to [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 682.832030] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1de30657-c45d-48c9-9f4e-98322e0bbc25 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.839674] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 682.839674] env[61985]: value = "task-935627" [ 682.839674] env[61985]: _type = "Task" [ 682.839674] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.848119] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935627, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.249646] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Releasing lock "refresh_cache-7f32151e-7be0-40e2-b210-72765b7551bc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.249870] env[61985]: DEBUG nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 683.250070] env[61985]: DEBUG nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.250245] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 683.272429] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.350405] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935627, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.398448] env[61985]: INFO nova.scheduler.client.report [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Deleted allocations for instance 9ac9f4b8-fcef-458a-913c-72cfc3322b6e [ 683.583309] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186ddf94-2012-40d4-8f6f-a771bfad0ea2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.590666] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad58b67-3dd0-4aab-98c0-791d37235e65 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.623320] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adb8897-fb9f-4094-b952-823300a863f5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.630726] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2edd37-55e7-4f5b-8fbe-e7b2419b8b7d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.643482] env[61985]: DEBUG nova.compute.provider_tree [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.774702] env[61985]: DEBUG nova.network.neutron [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.850915] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935627, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.662882} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.851203] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Copied Virtual Disk [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk to [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 683.851386] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleting the datastore file [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922/tmp-sparse.vmdk {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 683.851633] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81ee53b3-33e5-43d9-8cfc-f04dc2a9aec6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.857738] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 683.857738] env[61985]: value = "task-935628" [ 683.857738] env[61985]: _type = "Task" [ 683.857738] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.865366] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.909239] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f0c9ccf-397c-43ad-9b47-cf7f3af3b5c7 tempest-ServersWithSpecificFlavorTestJSON-1816748974 tempest-ServersWithSpecificFlavorTestJSON-1816748974-project-member] Lock "9ac9f4b8-fcef-458a-913c-72cfc3322b6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.569s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.146901] env[61985]: DEBUG nova.scheduler.client.report [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.280191] env[61985]: INFO nova.compute.manager [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] [instance: 7f32151e-7be0-40e2-b210-72765b7551bc] Took 1.03 seconds to deallocate network for instance. [ 684.367914] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023092} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.367914] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 684.368204] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Moving file from [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c/fe6d817b-6194-440f-988a-f4a94c580922 to [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922. {{(pid=61985) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 684.369088] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-00a85dbe-4fda-410d-a739-14a4180a2c54 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.376144] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 684.376144] env[61985]: value = "task-935629" [ 684.376144] env[61985]: _type = "Task" [ 684.376144] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.384725] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935629, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.411508] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 684.652835] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.653386] env[61985]: DEBUG nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 684.656478] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.886s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.657917] env[61985]: INFO nova.compute.claims [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.888109] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935629, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024945} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.888457] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] File moved {{(pid=61985) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 684.888720] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Cleaning up location [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 684.888957] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleting the datastore file [datastore1] vmware_temp/3433e2eb-008e-4551-986e-b6b36e81915c {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 684.889334] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-767d516e-4ab6-47e3-8165-8866c78d3626 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.896178] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 684.896178] env[61985]: value = "task-935630" [ 684.896178] env[61985]: _type = "Task" [ 684.896178] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.904114] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.941272] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.162255] env[61985]: DEBUG nova.compute.utils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.165474] env[61985]: DEBUG nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 685.165646] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 685.212607] env[61985]: DEBUG nova.policy [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a18774b772d47d69e5db077ded2f630', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32548f19ed1a4555872a929f20180d1f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 685.306480] env[61985]: INFO nova.scheduler.client.report [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Deleted allocations for instance 7f32151e-7be0-40e2-b210-72765b7551bc [ 685.410781] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935630, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024652} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.411222] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 685.412274] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06f09080-2b43-4c69-8666-c74102eb3047 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.419051] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 685.419051] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5241543a-73dc-d28a-54f4-0c84f1e4e5f6" [ 685.419051] env[61985]: _type = "Task" [ 685.419051] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.429486] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5241543a-73dc-d28a-54f4-0c84f1e4e5f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.665542] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Successfully created port: 79c4705e-03af-44b4-9b24-6f4562847269 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 685.668531] env[61985]: DEBUG nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 685.823028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1abf8b50-96c6-4234-8648-96dfda68e1a8 tempest-DeleteServersAdminTestJSON-1190591945 tempest-DeleteServersAdminTestJSON-1190591945-project-member] Lock "7f32151e-7be0-40e2-b210-72765b7551bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.066s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.932056] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5241543a-73dc-d28a-54f4-0c84f1e4e5f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009078} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.932384] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.932717] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 213a3e19-5589-4261-96b0-69acfb6798ef/213a3e19-5589-4261-96b0-69acfb6798ef.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 685.933063] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a722f800-7c3f-487f-aa9a-e10e2d122708 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.943498] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 685.943498] env[61985]: value = "task-935631" [ 685.943498] env[61985]: _type = "Task" [ 685.943498] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.951157] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935631, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.206044] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c107ca9-553a-4656-a5c9-2ebb247c20f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.212750] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c849d9-0378-4a46-afdd-7e212478e211 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.248949] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57a73c4-e062-4a0e-a7ce-6fc1d52fcf9b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.259357] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73833528-e4a6-4fcd-8b76-51fc16eb5ac5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.275723] env[61985]: DEBUG nova.compute.provider_tree [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.325248] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 686.453702] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935631, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478264} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.453702] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 213a3e19-5589-4261-96b0-69acfb6798ef/213a3e19-5589-4261-96b0-69acfb6798ef.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 686.453702] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 686.454703] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5b8768c-2edf-4376-9d06-9a497f343a8b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.460085] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 686.460085] env[61985]: value = "task-935632" [ 686.460085] env[61985]: _type = "Task" [ 686.460085] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.468358] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935632, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.684639] env[61985]: DEBUG nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 686.712984] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 686.713266] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 686.713424] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.713605] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 686.713756] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.713903] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 686.714145] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 686.714314] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 686.714486] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 686.714784] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 686.714973] env[61985]: DEBUG nova.virt.hardware [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.715840] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf2bd10-9eeb-49ef-9525-dda9ab90020a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.725100] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2b96a0-7546-446c-83a1-6f9e1e201879 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.779013] env[61985]: DEBUG nova.scheduler.client.report [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.848616] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.947162] env[61985]: DEBUG nova.compute.manager [req-ec81c60a-b8b4-4fc3-b055-e69e622cc798 req-eb7664f1-f501-4598-8b0e-4b5b2d6b213a service nova] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Received event network-changed-79c4705e-03af-44b4-9b24-6f4562847269 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 686.947162] env[61985]: DEBUG nova.compute.manager [req-ec81c60a-b8b4-4fc3-b055-e69e622cc798 req-eb7664f1-f501-4598-8b0e-4b5b2d6b213a service nova] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Refreshing instance network info cache due to event network-changed-79c4705e-03af-44b4-9b24-6f4562847269. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 686.947162] env[61985]: DEBUG oslo_concurrency.lockutils [req-ec81c60a-b8b4-4fc3-b055-e69e622cc798 req-eb7664f1-f501-4598-8b0e-4b5b2d6b213a service nova] Acquiring lock "refresh_cache-ba2fd184-e8cd-4667-8a41-eb1994c3329f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.947162] env[61985]: DEBUG oslo_concurrency.lockutils [req-ec81c60a-b8b4-4fc3-b055-e69e622cc798 req-eb7664f1-f501-4598-8b0e-4b5b2d6b213a service nova] Acquired lock "refresh_cache-ba2fd184-e8cd-4667-8a41-eb1994c3329f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.947162] env[61985]: DEBUG nova.network.neutron [req-ec81c60a-b8b4-4fc3-b055-e69e622cc798 req-eb7664f1-f501-4598-8b0e-4b5b2d6b213a service nova] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Refreshing network info cache for port 79c4705e-03af-44b4-9b24-6f4562847269 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 686.972696] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935632, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061881} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.973730] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 686.974689] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f47a5c-76f3-4c85-aa15-29dbad955100 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.997163] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Reconfiguring VM instance instance-00000018 to attach disk [datastore1] 213a3e19-5589-4261-96b0-69acfb6798ef/213a3e19-5589-4261-96b0-69acfb6798ef.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 686.997688] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a39715b1-a489-4978-881e-ed8f9740a438 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.018511] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 687.018511] env[61985]: value = "task-935633" [ 687.018511] env[61985]: _type = "Task" [ 687.018511] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.028555] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935633, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.286426] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.628s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.286426] env[61985]: DEBUG nova.compute.manager [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 687.291403] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.853s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.321263] env[61985]: ERROR nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 79c4705e-03af-44b4-9b24-6f4562847269, please check neutron logs for more information. [ 687.321263] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 687.321263] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.321263] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 687.321263] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.321263] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 687.321263] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.321263] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 687.321263] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.321263] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 687.321263] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.321263] env[61985]: ERROR nova.compute.manager raise self.value [ 687.321263] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.321263] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 687.321263] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.321263] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 687.321674] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.321674] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 687.321674] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 79c4705e-03af-44b4-9b24-6f4562847269, please check neutron logs for more information. [ 687.321674] env[61985]: ERROR nova.compute.manager [ 687.321674] env[61985]: Traceback (most recent call last): [ 687.321674] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 687.321674] env[61985]: listener.cb(fileno) [ 687.321674] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.321674] env[61985]: result = function(*args, **kwargs) [ 687.321674] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.321674] env[61985]: return func(*args, **kwargs) [ 687.321674] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.321674] env[61985]: raise e [ 687.321674] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.321674] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 687.321674] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.321674] env[61985]: created_port_ids = self._update_ports_for_instance( [ 687.321674] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.321674] env[61985]: with excutils.save_and_reraise_exception(): [ 687.321674] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.321674] env[61985]: self.force_reraise() [ 687.321674] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.321674] env[61985]: raise self.value [ 687.321674] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.321674] env[61985]: updated_port = self._update_port( [ 687.321674] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.321674] env[61985]: _ensure_no_port_binding_failure(port) [ 687.321674] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.321674] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 687.322367] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 79c4705e-03af-44b4-9b24-6f4562847269, please check neutron logs for more information. [ 687.322367] env[61985]: Removing descriptor: 17 [ 687.322367] env[61985]: ERROR nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 79c4705e-03af-44b4-9b24-6f4562847269, please check neutron logs for more information. [ 687.322367] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Traceback (most recent call last): [ 687.322367] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 687.322367] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] yield resources [ 687.322367] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 687.322367] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] self.driver.spawn(context, instance, image_meta, [ 687.322367] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 687.322367] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.322367] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.322367] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] vm_ref = self.build_virtual_machine(instance, [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] for vif in network_info: [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] return self._sync_wrapper(fn, *args, **kwargs) [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] self.wait() [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] self[:] = self._gt.wait() [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] return self._exit_event.wait() [ 687.322680] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] result = hub.switch() [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] return self.greenlet.switch() [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] result = function(*args, **kwargs) [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] return func(*args, **kwargs) [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] raise e [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] nwinfo = self.network_api.allocate_for_instance( [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.323013] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] created_port_ids = self._update_ports_for_instance( [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] with excutils.save_and_reraise_exception(): [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] self.force_reraise() [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] raise self.value [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] updated_port = self._update_port( [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] _ensure_no_port_binding_failure(port) [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.324473] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] raise exception.PortBindingFailed(port_id=port['id']) [ 687.324781] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] nova.exception.PortBindingFailed: Binding failed for port 79c4705e-03af-44b4-9b24-6f4562847269, please check neutron logs for more information. [ 687.324781] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] [ 687.324781] env[61985]: INFO nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Terminating instance [ 687.324865] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Acquiring lock "refresh_cache-ba2fd184-e8cd-4667-8a41-eb1994c3329f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.479478] env[61985]: DEBUG nova.network.neutron [req-ec81c60a-b8b4-4fc3-b055-e69e622cc798 req-eb7664f1-f501-4598-8b0e-4b5b2d6b213a service nova] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.529951] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935633, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.608947] env[61985]: DEBUG nova.network.neutron [req-ec81c60a-b8b4-4fc3-b055-e69e622cc798 req-eb7664f1-f501-4598-8b0e-4b5b2d6b213a service nova] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.801104] env[61985]: DEBUG nova.compute.utils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.805565] env[61985]: DEBUG nova.compute.manager [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Not allocating networking since 'none' was specified. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 688.029863] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935633, 'name': ReconfigVM_Task, 'duration_secs': 0.527541} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.031751] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Reconfigured VM instance instance-00000018 to attach disk [datastore1] 213a3e19-5589-4261-96b0-69acfb6798ef/213a3e19-5589-4261-96b0-69acfb6798ef.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 688.032518] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b62a049b-46ed-4551-9b73-186c1939bd8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.038678] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 688.038678] env[61985]: value = "task-935634" [ 688.038678] env[61985]: _type = "Task" [ 688.038678] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.048577] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935634, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.111697] env[61985]: DEBUG oslo_concurrency.lockutils [req-ec81c60a-b8b4-4fc3-b055-e69e622cc798 req-eb7664f1-f501-4598-8b0e-4b5b2d6b213a service nova] Releasing lock "refresh_cache-ba2fd184-e8cd-4667-8a41-eb1994c3329f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.112193] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Acquired lock "refresh_cache-ba2fd184-e8cd-4667-8a41-eb1994c3329f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.112387] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 688.215398] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60e2d56-ee08-4637-adc8-ee058ed25595 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.222912] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec3ac9d-9801-4ed0-b1ec-9cdb8addd9b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.254685] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d940d6b-a8ba-4f09-9340-f6bd73df1cd5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.261559] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf81f61-4b90-4a48-9b32-01931b8c2a32 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.275425] env[61985]: DEBUG nova.compute.provider_tree [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.306056] env[61985]: DEBUG nova.compute.manager [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 688.550023] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935634, 'name': Rename_Task, 'duration_secs': 0.167945} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.550023] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 688.550023] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-477daba6-8b59-4e77-8a8d-7daa91241254 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.556256] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 688.556256] env[61985]: value = "task-935635" [ 688.556256] env[61985]: _type = "Task" [ 688.556256] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.563838] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.642842] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.778688] env[61985]: DEBUG nova.scheduler.client.report [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.793327] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.988974] env[61985]: DEBUG nova.compute.manager [req-cad13971-cdef-4f9b-aed1-17238e54d216 req-b45d5774-92f8-42e3-b6ac-4f8e1360b1c9 service nova] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Received event network-vif-deleted-79c4705e-03af-44b4-9b24-6f4562847269 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 689.066691] env[61985]: DEBUG oslo_vmware.api [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935635, 'name': PowerOnVM_Task, 'duration_secs': 0.425383} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.066980] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 689.067207] env[61985]: INFO nova.compute.manager [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Took 8.82 seconds to spawn the instance on the hypervisor. [ 689.067308] env[61985]: DEBUG nova.compute.manager [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 689.068126] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2874a861-0da0-455f-9846-2aa69fe5bab5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.283636] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.284281] env[61985]: ERROR nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 892a5f88-22a3-44b3-8b77-8b48f536dd80, please check neutron logs for more information. [ 689.284281] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Traceback (most recent call last): [ 689.284281] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 689.284281] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] self.driver.spawn(context, instance, image_meta, [ 689.284281] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.284281] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.284281] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.284281] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] vm_ref = self.build_virtual_machine(instance, [ 689.284281] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.284281] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.284281] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] for vif in network_info: [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] return self._sync_wrapper(fn, *args, **kwargs) [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] self.wait() [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] self[:] = self._gt.wait() [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] return self._exit_event.wait() [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] result = hub.switch() [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.284634] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] return self.greenlet.switch() [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] result = function(*args, **kwargs) [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] return func(*args, **kwargs) [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] raise e [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] nwinfo = self.network_api.allocate_for_instance( [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] created_port_ids = self._update_ports_for_instance( [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] with excutils.save_and_reraise_exception(): [ 689.285052] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] self.force_reraise() [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] raise self.value [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] updated_port = self._update_port( [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] _ensure_no_port_binding_failure(port) [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] raise exception.PortBindingFailed(port_id=port['id']) [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] nova.exception.PortBindingFailed: Binding failed for port 892a5f88-22a3-44b3-8b77-8b48f536dd80, please check neutron logs for more information. [ 689.285411] env[61985]: ERROR nova.compute.manager [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] [ 689.285720] env[61985]: DEBUG nova.compute.utils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Binding failed for port 892a5f88-22a3-44b3-8b77-8b48f536dd80, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 689.286382] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.073s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.289592] env[61985]: DEBUG nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Build of instance 535a5c48-c816-44f7-a58c-734878d90cc7 was re-scheduled: Binding failed for port 892a5f88-22a3-44b3-8b77-8b48f536dd80, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 689.289765] env[61985]: DEBUG nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 689.289956] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "refresh_cache-535a5c48-c816-44f7-a58c-734878d90cc7" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.290114] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquired lock "refresh_cache-535a5c48-c816-44f7-a58c-734878d90cc7" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.290272] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 689.296717] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Releasing lock "refresh_cache-ba2fd184-e8cd-4667-8a41-eb1994c3329f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.297078] env[61985]: DEBUG nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 689.297625] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 689.297625] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-980fc210-ba40-4161-9771-09dcfeab7844 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.307668] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c0c308-b768-4f58-ad8e-9c9ee5a75df6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.320663] env[61985]: DEBUG nova.compute.manager [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 689.334834] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ba2fd184-e8cd-4667-8a41-eb1994c3329f could not be found. [ 689.335074] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 689.335258] env[61985]: INFO nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 689.335602] env[61985]: DEBUG oslo.service.loopingcall [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.335839] env[61985]: DEBUG nova.compute.manager [-] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.335952] env[61985]: DEBUG nova.network.neutron [-] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 689.345032] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.345272] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.345419] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.345605] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.345751] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.345897] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.346124] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.346286] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.346459] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.346636] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.346812] env[61985]: DEBUG nova.virt.hardware [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.347655] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9bb8ad-646b-46dc-86b0-04fa462b363a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.354597] env[61985]: DEBUG nova.network.neutron [-] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.357838] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f696eef-abd7-4716-bf61-b26676ba3792 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.372752] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 689.378286] env[61985]: DEBUG oslo.service.loopingcall [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.379034] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 689.379034] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bccc1721-aa73-437a-bf35-5c5588a82a5e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.395593] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 689.395593] env[61985]: value = "task-935636" [ 689.395593] env[61985]: _type = "Task" [ 689.395593] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.406212] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935636, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.589583] env[61985]: INFO nova.compute.manager [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Took 34.44 seconds to build instance. [ 689.830972] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.857693] env[61985]: DEBUG nova.network.neutron [-] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.910114] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935636, 'name': CreateVM_Task, 'duration_secs': 0.271434} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.910291] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 689.910707] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.910856] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.911187] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 689.911434] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d00e3607-2236-4eff-850e-6548710b8679 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.915905] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 689.915905] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523055f5-21e8-fe3d-d73f-0b6d5dedb5c3" [ 689.915905] env[61985]: _type = "Task" [ 689.915905] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.934999] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523055f5-21e8-fe3d-d73f-0b6d5dedb5c3, 'name': SearchDatastore_Task, 'duration_secs': 0.010222} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.937718] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.938045] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 689.938282] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.938437] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.938649] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 689.939417] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6e383a2d-2631-430c-b3ac-d3c19a6bc190 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.946696] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 689.946932] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 689.950076] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-150715e9-1079-49cf-ada6-e07f3e870780 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.956652] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 689.956652] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526a77d7-91fa-b487-d1a7-1b17359a6717" [ 689.956652] env[61985]: _type = "Task" [ 689.956652] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.964100] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526a77d7-91fa-b487-d1a7-1b17359a6717, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.036349] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.092315] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b41ae7fb-59bf-495c-832a-06db565b832e tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "213a3e19-5589-4261-96b0-69acfb6798ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.549s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.283045] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbeed21-95ab-4f65-b16d-a3fed6c344ae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.290127] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0507df11-4e9d-4d2b-b438-a6cc19ff8440 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.327884] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4babe39-29c4-4b33-ae9b-0515c52894b7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.335409] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e94dc42-7048-45d4-8a58-a53954893569 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.348797] env[61985]: DEBUG nova.compute.provider_tree [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.363790] env[61985]: INFO nova.compute.manager [-] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Took 1.03 seconds to deallocate network for instance. [ 690.365893] env[61985]: DEBUG nova.compute.claims [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 690.366092] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.467453] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526a77d7-91fa-b487-d1a7-1b17359a6717, 'name': SearchDatastore_Task, 'duration_secs': 0.008088} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.468264] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e875500-4886-491e-929e-d6fdca4e3239 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.474013] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 690.474013] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a66899-07fa-1f3d-559b-80da3f8cf5ac" [ 690.474013] env[61985]: _type = "Task" [ 690.474013] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.486095] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a66899-07fa-1f3d-559b-80da3f8cf5ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.542409] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Releasing lock "refresh_cache-535a5c48-c816-44f7-a58c-734878d90cc7" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.542409] env[61985]: DEBUG nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 690.542409] env[61985]: DEBUG nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.542409] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 690.565149] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.594789] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 690.852676] env[61985]: DEBUG nova.scheduler.client.report [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.987288] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a66899-07fa-1f3d-559b-80da3f8cf5ac, 'name': SearchDatastore_Task, 'duration_secs': 0.010043} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.987288] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.987288] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 2e1e46ea-477c-41f3-938f-bd654bc5999d/2e1e46ea-477c-41f3-938f-bd654bc5999d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 690.987807] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30b40197-38d8-4246-9c6f-0f420ce9057c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.999126] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 690.999126] env[61985]: value = "task-935637" [ 690.999126] env[61985]: _type = "Task" [ 690.999126] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.010423] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.067843] env[61985]: DEBUG nova.network.neutron [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.120149] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.358571] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.072s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.359254] env[61985]: ERROR nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a1fe1a7b-3745-42c6-9f72-0afc45077ddf, please check neutron logs for more information. [ 691.359254] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Traceback (most recent call last): [ 691.359254] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 691.359254] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] self.driver.spawn(context, instance, image_meta, [ 691.359254] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 691.359254] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.359254] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.359254] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] vm_ref = self.build_virtual_machine(instance, [ 691.359254] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.359254] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.359254] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] for vif in network_info: [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] return self._sync_wrapper(fn, *args, **kwargs) [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] self.wait() [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] self[:] = self._gt.wait() [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] return self._exit_event.wait() [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] result = hub.switch() [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 691.359815] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] return self.greenlet.switch() [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] result = function(*args, **kwargs) [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] return func(*args, **kwargs) [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] raise e [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] nwinfo = self.network_api.allocate_for_instance( [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] created_port_ids = self._update_ports_for_instance( [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] with excutils.save_and_reraise_exception(): [ 691.360189] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] self.force_reraise() [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] raise self.value [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] updated_port = self._update_port( [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] _ensure_no_port_binding_failure(port) [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] raise exception.PortBindingFailed(port_id=port['id']) [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] nova.exception.PortBindingFailed: Binding failed for port a1fe1a7b-3745-42c6-9f72-0afc45077ddf, please check neutron logs for more information. [ 691.360570] env[61985]: ERROR nova.compute.manager [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] [ 691.361495] env[61985]: DEBUG nova.compute.utils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Binding failed for port a1fe1a7b-3745-42c6-9f72-0afc45077ddf, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 691.361846] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.246s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.365649] env[61985]: DEBUG nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Build of instance c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124 was re-scheduled: Binding failed for port a1fe1a7b-3745-42c6-9f72-0afc45077ddf, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 691.366157] env[61985]: DEBUG nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 691.366402] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "refresh_cache-c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.366569] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "refresh_cache-c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.366789] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 691.509588] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935637, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510916} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.509871] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 2e1e46ea-477c-41f3-938f-bd654bc5999d/2e1e46ea-477c-41f3-938f-bd654bc5999d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 691.510986] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 691.510986] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f26f5882-752b-4064-bec6-7f9107d85547 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.518307] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 691.518307] env[61985]: value = "task-935638" [ 691.518307] env[61985]: _type = "Task" [ 691.518307] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.527103] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935638, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.572473] env[61985]: INFO nova.compute.manager [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 535a5c48-c816-44f7-a58c-734878d90cc7] Took 1.03 seconds to deallocate network for instance. [ 691.940478] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.031537] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935638, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067976} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.031801] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 692.033627] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd4cb32-d2b4-4426-aabc-ba8f0f514552 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.062898] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] 2e1e46ea-477c-41f3-938f-bd654bc5999d/2e1e46ea-477c-41f3-938f-bd654bc5999d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 692.068435] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d7227a6-77ed-49a2-9983-b5760d819654 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.097969] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 692.097969] env[61985]: value = "task-935639" [ 692.097969] env[61985]: _type = "Task" [ 692.097969] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.107643] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935639, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.199016] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.414034] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9650da8-9ee0-4bf9-8830-81fbaed76321 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.425248] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23faf2cd-9dde-41ad-bcbc-4dcc602b7382 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.458792] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5ee25e-e6eb-4d66-8745-92da0a581308 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.467318] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa7cc0b-2fe2-4925-8b86-c4e677f2e6a2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.482495] env[61985]: DEBUG nova.compute.provider_tree [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 692.612651] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935639, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.621834] env[61985]: INFO nova.scheduler.client.report [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Deleted allocations for instance 535a5c48-c816-44f7-a58c-734878d90cc7 [ 692.704849] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "refresh_cache-c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.704849] env[61985]: DEBUG nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 692.704849] env[61985]: DEBUG nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 692.704849] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 692.736287] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.008169] env[61985]: ERROR nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [req-5444c817-7e6c-4261-8b81-cbbe60d3db60] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5444c817-7e6c-4261-8b81-cbbe60d3db60"}]}: nova.exception.PortBindingFailed: Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. [ 693.034746] env[61985]: DEBUG nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 693.053823] env[61985]: DEBUG nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 693.054079] env[61985]: DEBUG nova.compute.provider_tree [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 55 to 56 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 693.054252] env[61985]: DEBUG nova.compute.provider_tree [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 693.067644] env[61985]: DEBUG nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 693.085957] env[61985]: DEBUG nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 693.112842] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935639, 'name': ReconfigVM_Task, 'duration_secs': 1.012499} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.113110] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Reconfigured VM instance instance-0000001a to attach disk [datastore2] 2e1e46ea-477c-41f3-938f-bd654bc5999d/2e1e46ea-477c-41f3-938f-bd654bc5999d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 693.113737] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7805c6be-d155-404c-8d23-b06ccb21f082 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.123329] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 693.123329] env[61985]: value = "task-935640" [ 693.123329] env[61985]: _type = "Task" [ 693.123329] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.131797] env[61985]: DEBUG oslo_concurrency.lockutils [None req-975114ca-92d9-4647-9764-007b5fea6fd3 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "535a5c48-c816-44f7-a58c-734878d90cc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.267s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.131797] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935640, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.240603] env[61985]: DEBUG nova.network.neutron [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.563663] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e4985a-fea5-4e75-9016-97d3e6f9c048 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.572952] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad74ad08-4c35-46a6-af74-3a79701ce61a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.582388] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquiring lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.582695] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.610494] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ef4629-301b-4c1f-acac-6cdcb0671d55 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.617977] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b367af83-709a-4497-8584-4008d78e3946 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.635810] env[61985]: DEBUG nova.compute.provider_tree [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 693.637179] env[61985]: DEBUG nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 693.643325] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935640, 'name': Rename_Task, 'duration_secs': 0.325971} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.643729] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 693.643957] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1effd4f7-b5fc-42b7-8375-4118db7cd9d7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.651020] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 693.651020] env[61985]: value = "task-935641" [ 693.651020] env[61985]: _type = "Task" [ 693.651020] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.660181] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935641, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.746029] env[61985]: INFO nova.compute.manager [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124] Took 1.04 seconds to deallocate network for instance. [ 694.164108] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935641, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.165031] env[61985]: ERROR nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [req-a1e5b089-3fac-48bb-8fbc-aa3c79b40100] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a1e5b089-3fac-48bb-8fbc-aa3c79b40100"}]}: nova.exception.PortBindingFailed: Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. [ 694.167951] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.179465] env[61985]: DEBUG nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 694.193566] env[61985]: DEBUG nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 694.193814] env[61985]: DEBUG nova.compute.provider_tree [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 694.204780] env[61985]: DEBUG nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 694.222913] env[61985]: DEBUG nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 694.441176] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquiring lock "402ef006-2835-4239-a3c4-1c37a7bba9a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.441421] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lock "402ef006-2835-4239-a3c4-1c37a7bba9a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.595192] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac9478c-9b54-4ad5-8bb2-c63ced19d24e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.603025] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a90cd7b-47d5-4bd4-9456-175bdf19a6c9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.633498] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203f6d42-1e4b-4f15-89ac-de6f45a0fb70 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.640195] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385bcf8e-5a86-4856-8791-a4af07a91849 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.652796] env[61985]: DEBUG nova.compute.provider_tree [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 694.661078] env[61985]: DEBUG oslo_vmware.api [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935641, 'name': PowerOnVM_Task, 'duration_secs': 0.762361} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.661903] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 694.662100] env[61985]: INFO nova.compute.manager [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Took 5.34 seconds to spawn the instance on the hypervisor. [ 694.662287] env[61985]: DEBUG nova.compute.manager [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 694.662979] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f10e3d-a550-4f2e-b585-d7b69f3892fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.776274] env[61985]: INFO nova.scheduler.client.report [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleted allocations for instance c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124 [ 695.179039] env[61985]: INFO nova.compute.manager [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Took 28.43 seconds to build instance. [ 695.183603] env[61985]: DEBUG nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 57 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 695.183851] env[61985]: DEBUG nova.compute.provider_tree [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 57 to 58 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 695.184043] env[61985]: DEBUG nova.compute.provider_tree [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 695.288046] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36c6c2de-e3f1-4783-969c-354268d49c3a tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "c2fe4f55-c1f8-46ea-ae6f-f0503d3ff124" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.164s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.680933] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cfdf3dde-4b21-44f6-ae4f-04808b8726c7 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "2e1e46ea-477c-41f3-938f-bd654bc5999d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.032s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.688383] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 4.326s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.689086] env[61985]: ERROR nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. [ 695.689086] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Traceback (most recent call last): [ 695.689086] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 695.689086] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] self.driver.spawn(context, instance, image_meta, [ 695.689086] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 695.689086] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.689086] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.689086] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] vm_ref = self.build_virtual_machine(instance, [ 695.689086] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.689086] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.689086] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] for vif in network_info: [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] return self._sync_wrapper(fn, *args, **kwargs) [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] self.wait() [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] self[:] = self._gt.wait() [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] return self._exit_event.wait() [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] result = hub.switch() [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 695.689410] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] return self.greenlet.switch() [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] result = function(*args, **kwargs) [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] return func(*args, **kwargs) [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] raise e [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] nwinfo = self.network_api.allocate_for_instance( [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] created_port_ids = self._update_ports_for_instance( [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] with excutils.save_and_reraise_exception(): [ 695.689852] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] self.force_reraise() [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] raise self.value [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] updated_port = self._update_port( [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] _ensure_no_port_binding_failure(port) [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] raise exception.PortBindingFailed(port_id=port['id']) [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] nova.exception.PortBindingFailed: Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. [ 695.690238] env[61985]: ERROR nova.compute.manager [instance: 955181b1-eb46-424c-8c79-055638dc9dac] [ 695.690568] env[61985]: DEBUG nova.compute.utils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 695.691403] env[61985]: DEBUG nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Build of instance 955181b1-eb46-424c-8c79-055638dc9dac was re-scheduled: Binding failed for port 2000fbab-b01f-4a68-bef7-1d5f32773dc1, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 695.692483] env[61985]: DEBUG nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 695.692900] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquiring lock "refresh_cache-955181b1-eb46-424c-8c79-055638dc9dac" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.692900] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Acquired lock "refresh_cache-955181b1-eb46-424c-8c79-055638dc9dac" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.693054] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 695.694501] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.149s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.790402] env[61985]: DEBUG nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 696.183785] env[61985]: DEBUG nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 696.222459] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.304113] env[61985]: INFO nova.compute.manager [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Rebuilding instance [ 696.318649] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.354623] env[61985]: DEBUG nova.compute.manager [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 696.358020] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4e7777-d91e-4ff2-8f7c-ea81b71522c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.386277] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.649195] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc42f6f-74b5-43d3-bc84-10a3869011d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.658281] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9c1e20-cdf5-411a-8cb2-57cfe36e7ca8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.692841] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b880d9b-90af-4168-9a8f-17948be11fc5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.701767] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aceec98-2ce5-4b31-8fd3-7b27369fc05a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.715981] env[61985]: DEBUG nova.compute.provider_tree [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.717554] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.869037] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 696.869311] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-635d1ac5-8b34-4825-98be-7d3136f5ef0c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.878365] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 696.878365] env[61985]: value = "task-935642" [ 696.878365] env[61985]: _type = "Task" [ 696.878365] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.886737] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935642, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.889353] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Releasing lock "refresh_cache-955181b1-eb46-424c-8c79-055638dc9dac" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.889611] env[61985]: DEBUG nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 696.889831] env[61985]: DEBUG nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.889994] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 696.910562] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.219167] env[61985]: DEBUG nova.scheduler.client.report [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.387860] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "c4e95f51-9acc-46e8-9921-e0a85fb38598" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.388461] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "c4e95f51-9acc-46e8-9921-e0a85fb38598" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.393192] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935642, 'name': PowerOffVM_Task, 'duration_secs': 0.120111} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.393448] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 697.393658] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 697.394420] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd134c2a-cd0d-4762-971f-7b8f4c08a012 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.401174] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 697.401390] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed9ba9c0-cdeb-48e9-8b08-de0d4cb4c26b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.412652] env[61985]: DEBUG nova.network.neutron [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.426387] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 697.426627] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 697.426817] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleting the datastore file [datastore2] 2e1e46ea-477c-41f3-938f-bd654bc5999d {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 697.427043] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49ded13e-1afe-420d-bc66-a5fb70a76139 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.432897] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 697.432897] env[61985]: value = "task-935644" [ 697.432897] env[61985]: _type = "Task" [ 697.432897] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.440554] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935644, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.723765] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.029s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.724402] env[61985]: ERROR nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f5e4273b-4de4-4a93-a78c-4615fa130904, please check neutron logs for more information. [ 697.724402] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Traceback (most recent call last): [ 697.724402] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 697.724402] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] self.driver.spawn(context, instance, image_meta, [ 697.724402] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 697.724402] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.724402] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.724402] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] vm_ref = self.build_virtual_machine(instance, [ 697.724402] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.724402] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.724402] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] for vif in network_info: [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] return self._sync_wrapper(fn, *args, **kwargs) [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] self.wait() [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] self[:] = self._gt.wait() [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] return self._exit_event.wait() [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] result = hub.switch() [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 697.724777] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] return self.greenlet.switch() [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] result = function(*args, **kwargs) [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] return func(*args, **kwargs) [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] raise e [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] nwinfo = self.network_api.allocate_for_instance( [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] created_port_ids = self._update_ports_for_instance( [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] with excutils.save_and_reraise_exception(): [ 697.725215] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] self.force_reraise() [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] raise self.value [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] updated_port = self._update_port( [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] _ensure_no_port_binding_failure(port) [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] raise exception.PortBindingFailed(port_id=port['id']) [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] nova.exception.PortBindingFailed: Binding failed for port f5e4273b-4de4-4a93-a78c-4615fa130904, please check neutron logs for more information. [ 697.725596] env[61985]: ERROR nova.compute.manager [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] [ 697.725913] env[61985]: DEBUG nova.compute.utils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Binding failed for port f5e4273b-4de4-4a93-a78c-4615fa130904, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 697.726338] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.478s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.728385] env[61985]: DEBUG nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Build of instance f86a5054-f30e-4868-9c50-2fe0d384b0b0 was re-scheduled: Binding failed for port f5e4273b-4de4-4a93-a78c-4615fa130904, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 697.728790] env[61985]: DEBUG nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 697.729030] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-f86a5054-f30e-4868-9c50-2fe0d384b0b0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.729175] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-f86a5054-f30e-4868-9c50-2fe0d384b0b0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.729340] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.915680] env[61985]: INFO nova.compute.manager [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] [instance: 955181b1-eb46-424c-8c79-055638dc9dac] Took 1.03 seconds to deallocate network for instance. [ 697.943597] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935644, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093141} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.944862] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 697.944862] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 697.944862] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 698.251143] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.360511] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.862636] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-f86a5054-f30e-4868-9c50-2fe0d384b0b0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.862891] env[61985]: DEBUG nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 698.863089] env[61985]: DEBUG nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.863260] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 698.877630] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.943408] env[61985]: INFO nova.scheduler.client.report [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Deleted allocations for instance 955181b1-eb46-424c-8c79-055638dc9dac [ 698.979034] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 698.979034] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 698.979034] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 698.979034] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 698.979333] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 698.979333] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 698.979422] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 698.979537] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 698.979712] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 698.979919] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 698.980141] env[61985]: DEBUG nova.virt.hardware [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 698.980991] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ddf579-1df5-49c0-a54a-0c9d2d9caf12 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.990482] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de2ad5bd-14ee-4dc6-9fac-624160e8a6b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.004926] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 699.010670] env[61985]: DEBUG oslo.service.loopingcall [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.010933] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 699.011200] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c2b7519-6bdd-4a37-a82b-603181e4f4c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.027926] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 699.027926] env[61985]: value = "task-935645" [ 699.027926] env[61985]: _type = "Task" [ 699.027926] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.035886] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935645, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.262951] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 955181b1-eb46-424c-8c79-055638dc9dac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.381278] env[61985]: DEBUG nova.network.neutron [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.456764] env[61985]: DEBUG oslo_concurrency.lockutils [None req-85db4f07-419e-4c3f-89c6-da02fcef90d5 tempest-ListServerFiltersTestJSON-1231520118 tempest-ListServerFiltersTestJSON-1231520118-project-member] Lock "955181b1-eb46-424c-8c79-055638dc9dac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.589s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.537778] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935645, 'name': CreateVM_Task, 'duration_secs': 0.251512} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.538155] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 699.538702] env[61985]: DEBUG oslo_concurrency.lockutils [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.538973] env[61985]: DEBUG oslo_concurrency.lockutils [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.539386] env[61985]: DEBUG oslo_concurrency.lockutils [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 699.541770] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bf0201a-de6f-49a2-be54-a7165beddfc5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.547026] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 699.547026] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5202bc26-4770-f624-4604-7f72b5b9130a" [ 699.547026] env[61985]: _type = "Task" [ 699.547026] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.552638] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5202bc26-4770-f624-4604-7f72b5b9130a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.767023] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance f86a5054-f30e-4868-9c50-2fe0d384b0b0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.767023] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 033766e8-92c0-498d-b3e8-37de5b46c20b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 699.767023] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 213a3e19-5589-4261-96b0-69acfb6798ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 699.767235] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance ba2fd184-e8cd-4667-8a41-eb1994c3329f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 699.767291] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 2e1e46ea-477c-41f3-938f-bd654bc5999d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 699.883939] env[61985]: INFO nova.compute.manager [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: f86a5054-f30e-4868-9c50-2fe0d384b0b0] Took 1.02 seconds to deallocate network for instance. [ 699.959176] env[61985]: DEBUG nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 700.058109] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5202bc26-4770-f624-4604-7f72b5b9130a, 'name': SearchDatastore_Task, 'duration_secs': 0.008815} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.058302] env[61985]: DEBUG oslo_concurrency.lockutils [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.058515] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 700.058794] env[61985]: DEBUG oslo_concurrency.lockutils [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.059104] env[61985]: DEBUG oslo_concurrency.lockutils [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.059196] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 700.059511] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-957aa1f6-b646-4017-be9c-473d0950e718 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.068169] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 700.068343] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 700.069038] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a69f2d90-b3a4-4ff8-9bb9-740e8e582148 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.073949] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 700.073949] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52725cbb-b070-055f-c346-c2ee1f5b2e15" [ 700.073949] env[61985]: _type = "Task" [ 700.073949] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.083936] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52725cbb-b070-055f-c346-c2ee1f5b2e15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.274399] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 40548796-fb8b-4b7e-9bfe-e83ca0edc01f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.478806] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.584869] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52725cbb-b070-055f-c346-c2ee1f5b2e15, 'name': SearchDatastore_Task, 'duration_secs': 0.008112} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.585668] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc787c1b-2966-4a42-98f4-2fa55b7babb0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.591613] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 700.591613] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52946cfa-e662-2b8b-e3eb-c083869617f4" [ 700.591613] env[61985]: _type = "Task" [ 700.591613] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.599740] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52946cfa-e662-2b8b-e3eb-c083869617f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.776553] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.914397] env[61985]: INFO nova.scheduler.client.report [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Deleted allocations for instance f86a5054-f30e-4868-9c50-2fe0d384b0b0 [ 701.102411] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52946cfa-e662-2b8b-e3eb-c083869617f4, 'name': SearchDatastore_Task, 'duration_secs': 0.008631} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.102703] env[61985]: DEBUG oslo_concurrency.lockutils [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.102966] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 2e1e46ea-477c-41f3-938f-bd654bc5999d/2e1e46ea-477c-41f3-938f-bd654bc5999d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 701.103261] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12e618fe-67b7-4b93-93ed-ecb493da546a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.110613] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 701.110613] env[61985]: value = "task-935646" [ 701.110613] env[61985]: _type = "Task" [ 701.110613] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.119545] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935646, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.279483] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 983c68ad-52ca-4eba-9570-3a37d8f5c637 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.424483] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8750674c-212d-429a-80c7-f1880d26b566 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "f86a5054-f30e-4868-9c50-2fe0d384b0b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.880s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.622984] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935646, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5097} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.623392] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 2e1e46ea-477c-41f3-938f-bd654bc5999d/2e1e46ea-477c-41f3-938f-bd654bc5999d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 701.623622] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 701.623909] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3531ffcd-0162-49f6-99a9-87bc94c7fac0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.630895] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 701.630895] env[61985]: value = "task-935647" [ 701.630895] env[61985]: _type = "Task" [ 701.630895] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.639253] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935647, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.784027] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance c8046008-3645-4845-be1e-23f5fefd5dc6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.927951] env[61985]: DEBUG nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 702.142523] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935647, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054975} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.142796] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 702.143615] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a2118f9-65a9-470d-9d2a-8f360dc85962 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.164432] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] 2e1e46ea-477c-41f3-938f-bd654bc5999d/2e1e46ea-477c-41f3-938f-bd654bc5999d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 702.164717] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fab086b1-9344-41f2-91a4-a8f7a1c0043a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.185043] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 702.185043] env[61985]: value = "task-935648" [ 702.185043] env[61985]: _type = "Task" [ 702.185043] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.193417] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935648, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.286285] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 3480f0ec-85f4-4881-ad17-d647cd401fbf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 702.450895] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.695813] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935648, 'name': ReconfigVM_Task, 'duration_secs': 0.273177} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.696125] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Reconfigured VM instance instance-0000001a to attach disk [datastore2] 2e1e46ea-477c-41f3-938f-bd654bc5999d/2e1e46ea-477c-41f3-938f-bd654bc5999d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 702.696795] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd041384-0dc6-4baf-93ff-2fa92c8dc1b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.704046] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 702.704046] env[61985]: value = "task-935649" [ 702.704046] env[61985]: _type = "Task" [ 702.704046] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.713042] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935649, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.794056] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 9af10e29-12ea-441d-b82a-be6976032c88 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.214657] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935649, 'name': Rename_Task, 'duration_secs': 0.124899} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.214930] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 703.215182] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c12b42f2-e1c0-4c49-bbed-c6f97f9d7659 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.222112] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 703.222112] env[61985]: value = "task-935650" [ 703.222112] env[61985]: _type = "Task" [ 703.222112] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.229598] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.297731] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 4a5bd6dd-9d21-43a2-9108-58aed0637ea9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.724101] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.724355] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.733918] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935650, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.801817] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance a0fd407f-7cf0-4c01-8195-399d8cbf62f9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.234481] env[61985]: DEBUG oslo_vmware.api [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935650, 'name': PowerOnVM_Task, 'duration_secs': 0.768478} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.234752] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 704.234950] env[61985]: DEBUG nova.compute.manager [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 704.235783] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c7fd57-6f44-4f30-9548-af881115b2d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.304263] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 50315640-7609-4fff-b191-aa29f06cb4f1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.750569] env[61985]: DEBUG oslo_concurrency.lockutils [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.808402] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 705.311168] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 2322ab72-9841-41fb-9d60-2812baabe108 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 705.355145] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "2e1e46ea-477c-41f3-938f-bd654bc5999d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.355432] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "2e1e46ea-477c-41f3-938f-bd654bc5999d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.355639] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "2e1e46ea-477c-41f3-938f-bd654bc5999d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.355820] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "2e1e46ea-477c-41f3-938f-bd654bc5999d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.355987] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "2e1e46ea-477c-41f3-938f-bd654bc5999d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.358352] env[61985]: INFO nova.compute.manager [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Terminating instance [ 705.359958] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "refresh_cache-2e1e46ea-477c-41f3-938f-bd654bc5999d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.360134] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired lock "refresh_cache-2e1e46ea-477c-41f3-938f-bd654bc5999d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.360305] env[61985]: DEBUG nova.network.neutron [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 705.814482] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 4add2608-8187-42a2-9897-3e5a9049757e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 705.883046] env[61985]: DEBUG nova.network.neutron [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.942508] env[61985]: DEBUG nova.network.neutron [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.317598] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance fea0cf39-e851-409f-86f5-31cc128a44dc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 706.445561] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Releasing lock "refresh_cache-2e1e46ea-477c-41f3-938f-bd654bc5999d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.446009] env[61985]: DEBUG nova.compute.manager [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 706.446235] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 706.447135] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549274fd-8dca-4b25-8591-5fcf41d9cf05 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.455711] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 706.455956] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26165050-d208-4e0e-a0a1-ab6fa9009e35 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.461913] env[61985]: DEBUG oslo_vmware.api [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 706.461913] env[61985]: value = "task-935651" [ 706.461913] env[61985]: _type = "Task" [ 706.461913] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.470119] env[61985]: DEBUG oslo_vmware.api [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935651, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.821119] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 794cff68-6e26-4607-96f7-eaeb41182551 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 706.972015] env[61985]: DEBUG oslo_vmware.api [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935651, 'name': PowerOffVM_Task, 'duration_secs': 0.154831} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.972291] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 706.972461] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 706.972702] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0388df87-422b-458f-909c-00ee57c0abc4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.996034] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 706.996034] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 706.996034] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleting the datastore file [datastore2] 2e1e46ea-477c-41f3-938f-bd654bc5999d {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 706.996236] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-704e018a-e83c-45c7-a6cf-4c643949c98d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.002025] env[61985]: DEBUG oslo_vmware.api [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 707.002025] env[61985]: value = "task-935653" [ 707.002025] env[61985]: _type = "Task" [ 707.002025] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.009207] env[61985]: DEBUG oslo_vmware.api [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935653, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.324086] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 3694e20c-ce37-4097-9991-8a06f38b2734 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 707.511616] env[61985]: DEBUG oslo_vmware.api [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935653, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089127} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.511864] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 707.512055] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 707.512230] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 707.512404] env[61985]: INFO nova.compute.manager [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Took 1.07 seconds to destroy the instance on the hypervisor. [ 707.512654] env[61985]: DEBUG oslo.service.loopingcall [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.514617] env[61985]: DEBUG nova.compute.manager [-] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.514617] env[61985]: DEBUG nova.network.neutron [-] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 707.526495] env[61985]: DEBUG nova.network.neutron [-] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.827062] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 3815af53-ba3b-4c60-a5fd-2d94102cdd42 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 708.029368] env[61985]: DEBUG nova.network.neutron [-] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.330506] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance f8bc8b71-0317-479d-b2f9-9471757f0774 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 708.535222] env[61985]: INFO nova.compute.manager [-] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Took 1.02 seconds to deallocate network for instance. [ 708.832296] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance b659ff28-5101-4825-84ea-111351c81145 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 709.043862] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.334877] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 761d5954-a5ca-4459-a1d6-bfc59b284bf4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 709.839076] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 4aca0ecb-4ae6-4400-accd-d71782b2806d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 710.341200] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance cfd59b61-cca9-48d5-85e1-1f45d13f1e88 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 710.844336] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8f6cd002-b3c2-4276-b195-15d09a143d31 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 711.348828] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance e6a57868-ab6e-45d2-9fa0-94efbb79185b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 711.851965] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 4a94a6f9-1c86-4628-aa63-341f2c114e2a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 712.354627] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 402ef006-2835-4239-a3c4-1c37a7bba9a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 712.858327] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance c4e95f51-9acc-46e8-9921-e0a85fb38598 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 712.858648] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 712.858838] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 713.197944] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f80ea77-ebd5-4c2e-b434-b6de9a39cc77 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.207179] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6341d3-2749-4421-bee6-76e64cffda7c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.236910] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9f2f68-1f9c-46d6-973a-a8f4f3724033 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.244216] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec68058-2cfc-4ff5-adb8-debefc19b7b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.256998] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.760212] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.265307] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 714.265577] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 16.539s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.265860] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 33.668s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.095962] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468e2d92-6553-4f7f-a782-6fcafefc8795 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.103546] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b587ce41-d22d-4148-96cb-4fd0491952c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.135078] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee47a20d-9ad2-4606-acda-6d571bda1c26 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.142912] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3174c30-5f34-4187-8640-d4f8d4501f83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.157065] env[61985]: DEBUG nova.compute.provider_tree [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.662016] env[61985]: DEBUG nova.scheduler.client.report [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.167366] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.901s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.168193] env[61985]: ERROR nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b2ee6e34-2ef1-4414-b847-2028086a3da9, please check neutron logs for more information. [ 716.168193] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Traceback (most recent call last): [ 716.168193] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 716.168193] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] self.driver.spawn(context, instance, image_meta, [ 716.168193] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 716.168193] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.168193] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.168193] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] vm_ref = self.build_virtual_machine(instance, [ 716.168193] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.168193] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.168193] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] for vif in network_info: [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] return self._sync_wrapper(fn, *args, **kwargs) [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] self.wait() [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] self[:] = self._gt.wait() [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] return self._exit_event.wait() [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] current.throw(*self._exc) [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.168658] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] result = function(*args, **kwargs) [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] return func(*args, **kwargs) [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] raise e [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] nwinfo = self.network_api.allocate_for_instance( [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] created_port_ids = self._update_ports_for_instance( [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] with excutils.save_and_reraise_exception(): [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] self.force_reraise() [ 716.169062] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.169410] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] raise self.value [ 716.169410] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 716.169410] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] updated_port = self._update_port( [ 716.169410] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.169410] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] _ensure_no_port_binding_failure(port) [ 716.169410] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.169410] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] raise exception.PortBindingFailed(port_id=port['id']) [ 716.169410] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] nova.exception.PortBindingFailed: Binding failed for port b2ee6e34-2ef1-4414-b847-2028086a3da9, please check neutron logs for more information. [ 716.169410] env[61985]: ERROR nova.compute.manager [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] [ 716.169410] env[61985]: DEBUG nova.compute.utils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Binding failed for port b2ee6e34-2ef1-4414-b847-2028086a3da9, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 716.170131] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.229s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.171590] env[61985]: INFO nova.compute.claims [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.174922] env[61985]: DEBUG nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Build of instance 033766e8-92c0-498d-b3e8-37de5b46c20b was re-scheduled: Binding failed for port b2ee6e34-2ef1-4414-b847-2028086a3da9, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 716.175388] env[61985]: DEBUG nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 716.175621] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Acquiring lock "refresh_cache-033766e8-92c0-498d-b3e8-37de5b46c20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.175774] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Acquired lock "refresh_cache-033766e8-92c0-498d-b3e8-37de5b46c20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.175937] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.695843] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.786127] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.288180] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Releasing lock "refresh_cache-033766e8-92c0-498d-b3e8-37de5b46c20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.288452] env[61985]: DEBUG nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 717.288579] env[61985]: DEBUG nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.288743] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 717.305309] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.504460] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63008bf2-f213-4d97-a76e-d6c7d683fd1b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.512068] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a220cc6d-c3d3-464f-9df5-3b2a6e4b1284 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.542908] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17342611-04b0-4c4a-aa92-05d252059484 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.550070] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09c41ec-8a83-4a3d-be13-81fc52ca7e1b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.563063] env[61985]: DEBUG nova.compute.provider_tree [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.807817] env[61985]: DEBUG nova.network.neutron [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.066535] env[61985]: DEBUG nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.311290] env[61985]: INFO nova.compute.manager [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] [instance: 033766e8-92c0-498d-b3e8-37de5b46c20b] Took 1.02 seconds to deallocate network for instance. [ 718.571997] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.572528] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 718.575711] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.727s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.577096] env[61985]: INFO nova.compute.claims [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.077609] env[61985]: DEBUG nova.compute.utils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.079059] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.079242] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 719.148864] env[61985]: DEBUG nova.policy [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b673311a9e2743d799f7a5b8649c7bcd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74df6d94aa0044f894f0c52ba53ec5f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.339106] env[61985]: INFO nova.scheduler.client.report [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Deleted allocations for instance 033766e8-92c0-498d-b3e8-37de5b46c20b [ 719.483940] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Successfully created port: b51b28a0-1c84-4229-895b-4546f9a0c449 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.584199] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 719.849157] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cda2be8f-ec88-4a64-b773-52ab766bde4a tempest-TenantUsagesTestJSON-1093953421 tempest-TenantUsagesTestJSON-1093953421-project-member] Lock "033766e8-92c0-498d-b3e8-37de5b46c20b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.107s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.080960] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c8a744-c63d-4048-aaf1-02c4fd376f8a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.088926] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5b2b78-305b-4942-8329-ce1e32107c77 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.127537] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813d7578-f038-4d5d-92ad-9aa3660617b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.135805] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ad42be-55a8-4291-9536-1782400d246a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.152761] env[61985]: DEBUG nova.compute.provider_tree [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.352185] env[61985]: DEBUG nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 720.533425] env[61985]: DEBUG nova.compute.manager [req-6b052d7a-c75d-493f-ada1-e94bc17516ae req-c9d07b08-edb3-4c5f-9cb8-2477e34b8217 service nova] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Received event network-changed-b51b28a0-1c84-4229-895b-4546f9a0c449 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 720.533639] env[61985]: DEBUG nova.compute.manager [req-6b052d7a-c75d-493f-ada1-e94bc17516ae req-c9d07b08-edb3-4c5f-9cb8-2477e34b8217 service nova] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Refreshing instance network info cache due to event network-changed-b51b28a0-1c84-4229-895b-4546f9a0c449. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 720.533861] env[61985]: DEBUG oslo_concurrency.lockutils [req-6b052d7a-c75d-493f-ada1-e94bc17516ae req-c9d07b08-edb3-4c5f-9cb8-2477e34b8217 service nova] Acquiring lock "refresh_cache-40548796-fb8b-4b7e-9bfe-e83ca0edc01f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.534014] env[61985]: DEBUG oslo_concurrency.lockutils [req-6b052d7a-c75d-493f-ada1-e94bc17516ae req-c9d07b08-edb3-4c5f-9cb8-2477e34b8217 service nova] Acquired lock "refresh_cache-40548796-fb8b-4b7e-9bfe-e83ca0edc01f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.534192] env[61985]: DEBUG nova.network.neutron [req-6b052d7a-c75d-493f-ada1-e94bc17516ae req-c9d07b08-edb3-4c5f-9cb8-2477e34b8217 service nova] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Refreshing network info cache for port b51b28a0-1c84-4229-895b-4546f9a0c449 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 720.598496] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 720.626965] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.627247] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.627405] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.627604] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.627752] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.627896] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.628117] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.628282] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.628473] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.628614] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.628787] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.629657] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58a4845-1a54-4333-8f4c-bb87ab0ce06e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.637588] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfceffb-db0c-4a5e-b03e-2ba2530bba51 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.657077] env[61985]: DEBUG nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.747324] env[61985]: ERROR nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b51b28a0-1c84-4229-895b-4546f9a0c449, please check neutron logs for more information. [ 720.747324] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 720.747324] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.747324] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 720.747324] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 720.747324] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 720.747324] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 720.747324] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 720.747324] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.747324] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 720.747324] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.747324] env[61985]: ERROR nova.compute.manager raise self.value [ 720.747324] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 720.747324] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 720.747324] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.747324] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 720.747858] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.747858] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 720.747858] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b51b28a0-1c84-4229-895b-4546f9a0c449, please check neutron logs for more information. [ 720.747858] env[61985]: ERROR nova.compute.manager [ 720.747858] env[61985]: Traceback (most recent call last): [ 720.747858] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 720.747858] env[61985]: listener.cb(fileno) [ 720.747858] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.747858] env[61985]: result = function(*args, **kwargs) [ 720.747858] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.747858] env[61985]: return func(*args, **kwargs) [ 720.747858] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.747858] env[61985]: raise e [ 720.747858] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.747858] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 720.747858] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 720.747858] env[61985]: created_port_ids = self._update_ports_for_instance( [ 720.747858] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 720.747858] env[61985]: with excutils.save_and_reraise_exception(): [ 720.747858] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.747858] env[61985]: self.force_reraise() [ 720.747858] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.747858] env[61985]: raise self.value [ 720.747858] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 720.747858] env[61985]: updated_port = self._update_port( [ 720.747858] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.747858] env[61985]: _ensure_no_port_binding_failure(port) [ 720.747858] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.747858] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 720.748655] env[61985]: nova.exception.PortBindingFailed: Binding failed for port b51b28a0-1c84-4229-895b-4546f9a0c449, please check neutron logs for more information. [ 720.748655] env[61985]: Removing descriptor: 17 [ 720.748655] env[61985]: ERROR nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b51b28a0-1c84-4229-895b-4546f9a0c449, please check neutron logs for more information. [ 720.748655] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Traceback (most recent call last): [ 720.748655] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 720.748655] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] yield resources [ 720.748655] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 720.748655] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] self.driver.spawn(context, instance, image_meta, [ 720.748655] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 720.748655] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.748655] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.748655] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] vm_ref = self.build_virtual_machine(instance, [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] for vif in network_info: [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] return self._sync_wrapper(fn, *args, **kwargs) [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] self.wait() [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] self[:] = self._gt.wait() [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] return self._exit_event.wait() [ 720.748954] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] result = hub.switch() [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] return self.greenlet.switch() [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] result = function(*args, **kwargs) [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] return func(*args, **kwargs) [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] raise e [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] nwinfo = self.network_api.allocate_for_instance( [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 720.749294] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] created_port_ids = self._update_ports_for_instance( [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] with excutils.save_and_reraise_exception(): [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] self.force_reraise() [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] raise self.value [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] updated_port = self._update_port( [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] _ensure_no_port_binding_failure(port) [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.749662] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] raise exception.PortBindingFailed(port_id=port['id']) [ 720.749935] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] nova.exception.PortBindingFailed: Binding failed for port b51b28a0-1c84-4229-895b-4546f9a0c449, please check neutron logs for more information. [ 720.749935] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] [ 720.749935] env[61985]: INFO nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Terminating instance [ 720.751230] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "refresh_cache-40548796-fb8b-4b7e-9bfe-e83ca0edc01f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.874375] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.052926] env[61985]: DEBUG nova.network.neutron [req-6b052d7a-c75d-493f-ada1-e94bc17516ae req-c9d07b08-edb3-4c5f-9cb8-2477e34b8217 service nova] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.160998] env[61985]: DEBUG nova.network.neutron [req-6b052d7a-c75d-493f-ada1-e94bc17516ae req-c9d07b08-edb3-4c5f-9cb8-2477e34b8217 service nova] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.163026] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.163556] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 721.165636] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.800s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.666744] env[61985]: DEBUG oslo_concurrency.lockutils [req-6b052d7a-c75d-493f-ada1-e94bc17516ae req-c9d07b08-edb3-4c5f-9cb8-2477e34b8217 service nova] Releasing lock "refresh_cache-40548796-fb8b-4b7e-9bfe-e83ca0edc01f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.669833] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquired lock "refresh_cache-40548796-fb8b-4b7e-9bfe-e83ca0edc01f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.669833] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 721.674780] env[61985]: DEBUG nova.compute.utils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.675730] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.676045] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 721.742799] env[61985]: DEBUG nova.policy [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b673311a9e2743d799f7a5b8649c7bcd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74df6d94aa0044f894f0c52ba53ec5f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 722.127371] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd0a930-9fc0-4759-a2ab-c19e729a947e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.131105] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Successfully created port: e51de715-fd1a-49a0-9cc4-4911cc14b0ec {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.138128] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8dbf7e1-8c45-4104-9740-b1f13cf93125 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.175319] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525e807c-b0ce-4572-8f51-5cbe564bf5d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.180557] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 722.191446] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ff692a-65cd-43dd-b7fd-a1a2a7f646f3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.203069] env[61985]: DEBUG nova.compute.provider_tree [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.216027] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.423721] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.587268] env[61985]: DEBUG nova.compute.manager [req-0bf72d17-1925-469d-9dde-0fac291d1604 req-b51ec7d5-43bc-424c-8bec-58ec1e1e0779 service nova] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Received event network-vif-deleted-b51b28a0-1c84-4229-895b-4546f9a0c449 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 722.709666] env[61985]: DEBUG nova.scheduler.client.report [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.929033] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Releasing lock "refresh_cache-40548796-fb8b-4b7e-9bfe-e83ca0edc01f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.929033] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 722.929033] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 722.929033] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a438394f-3a4f-4979-84dc-b615f0ff4123 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.942699] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bff9e94-6f33-48d2-bb34-f7abc7cbeb3c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.974408] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 40548796-fb8b-4b7e-9bfe-e83ca0edc01f could not be found. [ 722.974740] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 722.974995] env[61985]: INFO nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 722.975339] env[61985]: DEBUG oslo.service.loopingcall [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.975703] env[61985]: DEBUG nova.compute.manager [-] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.975808] env[61985]: DEBUG nova.network.neutron [-] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 722.998988] env[61985]: DEBUG nova.network.neutron [-] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.196043] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 723.216566] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.051s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.217364] env[61985]: ERROR nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 79c4705e-03af-44b4-9b24-6f4562847269, please check neutron logs for more information. [ 723.217364] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Traceback (most recent call last): [ 723.217364] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 723.217364] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] self.driver.spawn(context, instance, image_meta, [ 723.217364] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 723.217364] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.217364] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.217364] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] vm_ref = self.build_virtual_machine(instance, [ 723.217364] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.217364] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.217364] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] for vif in network_info: [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] return self._sync_wrapper(fn, *args, **kwargs) [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] self.wait() [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] self[:] = self._gt.wait() [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] return self._exit_event.wait() [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] result = hub.switch() [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 723.217747] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] return self.greenlet.switch() [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] result = function(*args, **kwargs) [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] return func(*args, **kwargs) [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] raise e [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] nwinfo = self.network_api.allocate_for_instance( [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] created_port_ids = self._update_ports_for_instance( [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] with excutils.save_and_reraise_exception(): [ 723.218196] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] self.force_reraise() [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] raise self.value [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] updated_port = self._update_port( [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] _ensure_no_port_binding_failure(port) [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] raise exception.PortBindingFailed(port_id=port['id']) [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] nova.exception.PortBindingFailed: Binding failed for port 79c4705e-03af-44b4-9b24-6f4562847269, please check neutron logs for more information. [ 723.218597] env[61985]: ERROR nova.compute.manager [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] [ 723.218964] env[61985]: DEBUG nova.compute.utils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Binding failed for port 79c4705e-03af-44b4-9b24-6f4562847269, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 723.219927] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.100s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.221414] env[61985]: INFO nova.compute.claims [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.224666] env[61985]: DEBUG nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Build of instance ba2fd184-e8cd-4667-8a41-eb1994c3329f was re-scheduled: Binding failed for port 79c4705e-03af-44b4-9b24-6f4562847269, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 723.225124] env[61985]: DEBUG nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 723.225401] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Acquiring lock "refresh_cache-ba2fd184-e8cd-4667-8a41-eb1994c3329f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.225577] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Acquired lock "refresh_cache-ba2fd184-e8cd-4667-8a41-eb1994c3329f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.225718] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.229366] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.229366] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.229366] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.229519] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.229519] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.229614] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.229825] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.229982] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.230165] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.230328] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.230495] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.231825] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e24632e-7391-4ea3-b1ed-397162a41bba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.241398] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e35dae-fda3-4688-8a82-cb6b5e99c1e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.501823] env[61985]: DEBUG nova.network.neutron [-] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.623982] env[61985]: ERROR nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. [ 723.623982] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 723.623982] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.623982] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 723.623982] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.623982] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 723.623982] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.623982] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 723.623982] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.623982] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 723.623982] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.623982] env[61985]: ERROR nova.compute.manager raise self.value [ 723.623982] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.623982] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 723.623982] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.623982] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 723.624467] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.624467] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 723.624467] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. [ 723.624467] env[61985]: ERROR nova.compute.manager [ 723.624467] env[61985]: Traceback (most recent call last): [ 723.624467] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 723.624467] env[61985]: listener.cb(fileno) [ 723.624467] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.624467] env[61985]: result = function(*args, **kwargs) [ 723.624467] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.624467] env[61985]: return func(*args, **kwargs) [ 723.624467] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.624467] env[61985]: raise e [ 723.624467] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.624467] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 723.624467] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.624467] env[61985]: created_port_ids = self._update_ports_for_instance( [ 723.624467] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.624467] env[61985]: with excutils.save_and_reraise_exception(): [ 723.624467] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.624467] env[61985]: self.force_reraise() [ 723.624467] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.624467] env[61985]: raise self.value [ 723.624467] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.624467] env[61985]: updated_port = self._update_port( [ 723.624467] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.624467] env[61985]: _ensure_no_port_binding_failure(port) [ 723.624467] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.624467] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 723.625269] env[61985]: nova.exception.PortBindingFailed: Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. [ 723.625269] env[61985]: Removing descriptor: 17 [ 723.625269] env[61985]: ERROR nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. [ 723.625269] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Traceback (most recent call last): [ 723.625269] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 723.625269] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] yield resources [ 723.625269] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 723.625269] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] self.driver.spawn(context, instance, image_meta, [ 723.625269] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 723.625269] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.625269] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.625269] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] vm_ref = self.build_virtual_machine(instance, [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] for vif in network_info: [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] return self._sync_wrapper(fn, *args, **kwargs) [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] self.wait() [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] self[:] = self._gt.wait() [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] return self._exit_event.wait() [ 723.625860] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] result = hub.switch() [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] return self.greenlet.switch() [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] result = function(*args, **kwargs) [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] return func(*args, **kwargs) [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] raise e [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] nwinfo = self.network_api.allocate_for_instance( [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.626305] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] created_port_ids = self._update_ports_for_instance( [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] with excutils.save_and_reraise_exception(): [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] self.force_reraise() [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] raise self.value [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] updated_port = self._update_port( [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] _ensure_no_port_binding_failure(port) [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.626671] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] raise exception.PortBindingFailed(port_id=port['id']) [ 723.626979] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] nova.exception.PortBindingFailed: Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. [ 723.626979] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] [ 723.626979] env[61985]: INFO nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Terminating instance [ 723.629599] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "refresh_cache-6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.632929] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquired lock "refresh_cache-6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.633108] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.757088] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.897492] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.004489] env[61985]: INFO nova.compute.manager [-] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Took 1.03 seconds to deallocate network for instance. [ 724.006936] env[61985]: DEBUG nova.compute.claims [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 724.007073] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.178490] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.484283] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Releasing lock "refresh_cache-ba2fd184-e8cd-4667-8a41-eb1994c3329f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.484283] env[61985]: DEBUG nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 724.484283] env[61985]: DEBUG nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.484283] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 724.484283] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.625175] env[61985]: DEBUG nova.compute.manager [req-d7a48921-4ef4-4c8a-9e96-027aecae2828 req-6581c099-9831-497e-9c61-87ca332a4150 service nova] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Received event network-changed-e51de715-fd1a-49a0-9cc4-4911cc14b0ec {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 724.625345] env[61985]: DEBUG nova.compute.manager [req-d7a48921-4ef4-4c8a-9e96-027aecae2828 req-6581c099-9831-497e-9c61-87ca332a4150 service nova] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Refreshing instance network info cache due to event network-changed-e51de715-fd1a-49a0-9cc4-4911cc14b0ec. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 724.625599] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7a48921-4ef4-4c8a-9e96-027aecae2828 req-6581c099-9831-497e-9c61-87ca332a4150 service nova] Acquiring lock "refresh_cache-6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.637769] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.660734] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.660997] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.901640] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5b6924-6556-4741-9f65-9772595f8721 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.910090] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92729471-3f19-441c-9bdc-a5d35a56b0a5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.946950] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd17c89-8eb7-4fdc-a056-ab53a43a55cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.954944] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3887db65-78b9-4227-aa92-336412895bcc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.969446] env[61985]: DEBUG nova.compute.provider_tree [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.986117] env[61985]: DEBUG nova.network.neutron [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.143120] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Releasing lock "refresh_cache-6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.143120] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 725.143120] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 725.143120] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7a48921-4ef4-4c8a-9e96-027aecae2828 req-6581c099-9831-497e-9c61-87ca332a4150 service nova] Acquired lock "refresh_cache-6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.143120] env[61985]: DEBUG nova.network.neutron [req-d7a48921-4ef4-4c8a-9e96-027aecae2828 req-6581c099-9831-497e-9c61-87ca332a4150 service nova] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Refreshing network info cache for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.143588] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0cda266e-7481-4c91-9317-f53c696b8eea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.154301] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9f12ab-5196-4120-8352-72461dde0e34 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.177131] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d could not be found. [ 725.177363] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 725.177582] env[61985]: INFO nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 725.177899] env[61985]: DEBUG oslo.service.loopingcall [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.178155] env[61985]: DEBUG nova.compute.manager [-] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.178249] env[61985]: DEBUG nova.network.neutron [-] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 725.202132] env[61985]: DEBUG nova.network.neutron [-] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.473206] env[61985]: DEBUG nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.487416] env[61985]: INFO nova.compute.manager [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] [instance: ba2fd184-e8cd-4667-8a41-eb1994c3329f] Took 1.09 seconds to deallocate network for instance. [ 725.663993] env[61985]: DEBUG nova.network.neutron [req-d7a48921-4ef4-4c8a-9e96-027aecae2828 req-6581c099-9831-497e-9c61-87ca332a4150 service nova] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.704417] env[61985]: DEBUG nova.network.neutron [-] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.774089] env[61985]: DEBUG nova.network.neutron [req-d7a48921-4ef4-4c8a-9e96-027aecae2828 req-6581c099-9831-497e-9c61-87ca332a4150 service nova] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.977669] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.758s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.978273] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 725.980892] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.813s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.982441] env[61985]: INFO nova.compute.claims [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.208801] env[61985]: INFO nova.compute.manager [-] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Took 1.03 seconds to deallocate network for instance. [ 726.211270] env[61985]: DEBUG nova.compute.claims [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 726.211270] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.277101] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7a48921-4ef4-4c8a-9e96-027aecae2828 req-6581c099-9831-497e-9c61-87ca332a4150 service nova] Releasing lock "refresh_cache-6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.277393] env[61985]: DEBUG nova.compute.manager [req-d7a48921-4ef4-4c8a-9e96-027aecae2828 req-6581c099-9831-497e-9c61-87ca332a4150 service nova] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Received event network-vif-deleted-e51de715-fd1a-49a0-9cc4-4911cc14b0ec {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 726.486819] env[61985]: DEBUG nova.compute.utils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.491104] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.491195] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 726.515413] env[61985]: INFO nova.scheduler.client.report [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Deleted allocations for instance ba2fd184-e8cd-4667-8a41-eb1994c3329f [ 726.529492] env[61985]: DEBUG nova.policy [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b673311a9e2743d799f7a5b8649c7bcd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74df6d94aa0044f894f0c52ba53ec5f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.888435] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Successfully created port: 6442d848-93a4-4438-a8bd-534aca1b64f5 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 726.992456] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 727.026246] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d4bd5e-2e2d-4452-8f34-0e31a94f0c53 tempest-AttachInterfacesUnderV243Test-772518966 tempest-AttachInterfacesUnderV243Test-772518966-project-member] Lock "ba2fd184-e8cd-4667-8a41-eb1994c3329f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.360s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.451689] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75671d54-2fed-4aa2-82fe-372dba7a3f1c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.459633] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e83ac0e-adf0-433d-8829-268279fe21b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.493436] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092aad08-4070-4d0c-8b92-0e7d124f5219 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.504749] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac05d0d-9e0e-40e6-aeae-5e5df1aeb2d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.520693] env[61985]: DEBUG nova.compute.provider_tree [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.531210] env[61985]: DEBUG nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 727.841367] env[61985]: DEBUG nova.compute.manager [req-d68cddf7-a963-4f83-8133-18a1fdc99763 req-bb85e43c-fc52-44c1-8fc4-bb0d708b428b service nova] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Received event network-changed-6442d848-93a4-4438-a8bd-534aca1b64f5 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 727.841581] env[61985]: DEBUG nova.compute.manager [req-d68cddf7-a963-4f83-8133-18a1fdc99763 req-bb85e43c-fc52-44c1-8fc4-bb0d708b428b service nova] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Refreshing instance network info cache due to event network-changed-6442d848-93a4-4438-a8bd-534aca1b64f5. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 727.841793] env[61985]: DEBUG oslo_concurrency.lockutils [req-d68cddf7-a963-4f83-8133-18a1fdc99763 req-bb85e43c-fc52-44c1-8fc4-bb0d708b428b service nova] Acquiring lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.841938] env[61985]: DEBUG oslo_concurrency.lockutils [req-d68cddf7-a963-4f83-8133-18a1fdc99763 req-bb85e43c-fc52-44c1-8fc4-bb0d708b428b service nova] Acquired lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.842615] env[61985]: DEBUG nova.network.neutron [req-d68cddf7-a963-4f83-8133-18a1fdc99763 req-bb85e43c-fc52-44c1-8fc4-bb0d708b428b service nova] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Refreshing network info cache for port 6442d848-93a4-4438-a8bd-534aca1b64f5 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 728.003891] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 728.026025] env[61985]: DEBUG nova.scheduler.client.report [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.031524] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.031781] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.032108] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.032351] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.032505] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.032655] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.032857] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.033190] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.033381] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.033550] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.033726] env[61985]: DEBUG nova.virt.hardware [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.034651] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38cfa9c-b78a-4517-b91d-0410e10e604d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.046473] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b3598d-5f8f-43c0-9e95-197d08acb8fb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.062765] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.076521] env[61985]: ERROR nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6442d848-93a4-4438-a8bd-534aca1b64f5, please check neutron logs for more information. [ 728.076521] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 728.076521] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.076521] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 728.076521] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.076521] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 728.076521] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.076521] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 728.076521] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.076521] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 728.076521] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.076521] env[61985]: ERROR nova.compute.manager raise self.value [ 728.076521] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.076521] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 728.076521] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.076521] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 728.077031] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.077031] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 728.077031] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6442d848-93a4-4438-a8bd-534aca1b64f5, please check neutron logs for more information. [ 728.077031] env[61985]: ERROR nova.compute.manager [ 728.077031] env[61985]: Traceback (most recent call last): [ 728.077031] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 728.077031] env[61985]: listener.cb(fileno) [ 728.077031] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.077031] env[61985]: result = function(*args, **kwargs) [ 728.077031] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.077031] env[61985]: return func(*args, **kwargs) [ 728.077031] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.077031] env[61985]: raise e [ 728.077031] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.077031] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 728.077031] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.077031] env[61985]: created_port_ids = self._update_ports_for_instance( [ 728.077031] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.077031] env[61985]: with excutils.save_and_reraise_exception(): [ 728.077031] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.077031] env[61985]: self.force_reraise() [ 728.077031] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.077031] env[61985]: raise self.value [ 728.077031] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.077031] env[61985]: updated_port = self._update_port( [ 728.077031] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.077031] env[61985]: _ensure_no_port_binding_failure(port) [ 728.077031] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.077031] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 728.077873] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 6442d848-93a4-4438-a8bd-534aca1b64f5, please check neutron logs for more information. [ 728.077873] env[61985]: Removing descriptor: 17 [ 728.077873] env[61985]: ERROR nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6442d848-93a4-4438-a8bd-534aca1b64f5, please check neutron logs for more information. [ 728.077873] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Traceback (most recent call last): [ 728.077873] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 728.077873] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] yield resources [ 728.077873] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 728.077873] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] self.driver.spawn(context, instance, image_meta, [ 728.077873] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 728.077873] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.077873] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.077873] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] vm_ref = self.build_virtual_machine(instance, [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] for vif in network_info: [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] return self._sync_wrapper(fn, *args, **kwargs) [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] self.wait() [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] self[:] = self._gt.wait() [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] return self._exit_event.wait() [ 728.078194] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] result = hub.switch() [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] return self.greenlet.switch() [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] result = function(*args, **kwargs) [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] return func(*args, **kwargs) [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] raise e [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] nwinfo = self.network_api.allocate_for_instance( [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.078568] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] created_port_ids = self._update_ports_for_instance( [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] with excutils.save_and_reraise_exception(): [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] self.force_reraise() [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] raise self.value [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] updated_port = self._update_port( [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] _ensure_no_port_binding_failure(port) [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.078943] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] raise exception.PortBindingFailed(port_id=port['id']) [ 728.079300] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] nova.exception.PortBindingFailed: Binding failed for port 6442d848-93a4-4438-a8bd-534aca1b64f5, please check neutron logs for more information. [ 728.079300] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] [ 728.079300] env[61985]: INFO nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Terminating instance [ 728.079627] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.374487] env[61985]: DEBUG nova.network.neutron [req-d68cddf7-a963-4f83-8133-18a1fdc99763 req-bb85e43c-fc52-44c1-8fc4-bb0d708b428b service nova] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.539782] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.540201] env[61985]: DEBUG nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 728.544107] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.225s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.545737] env[61985]: INFO nova.compute.claims [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.568019] env[61985]: DEBUG nova.network.neutron [req-d68cddf7-a963-4f83-8133-18a1fdc99763 req-bb85e43c-fc52-44c1-8fc4-bb0d708b428b service nova] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.045947] env[61985]: DEBUG nova.compute.utils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.047358] env[61985]: DEBUG nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 729.047727] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 729.071797] env[61985]: DEBUG oslo_concurrency.lockutils [req-d68cddf7-a963-4f83-8133-18a1fdc99763 req-bb85e43c-fc52-44c1-8fc4-bb0d708b428b service nova] Releasing lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.072187] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquired lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.072370] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 729.103865] env[61985]: DEBUG nova.policy [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f832e586e8b744f89388329f518ed6df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83e16978f0db4a3683dcdaf82655fc73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.551684] env[61985]: DEBUG nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 729.689555] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.693170] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Successfully created port: 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 729.877951] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.938796] env[61985]: DEBUG nova.compute.manager [req-85a218dd-63bc-4889-bdcf-317624be6f39 req-0156bee9-b24c-4e2f-bde5-e5dfa9015bfe service nova] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Received event network-vif-deleted-6442d848-93a4-4438-a8bd-534aca1b64f5 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 729.948581] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fadc90-b3ea-4f62-be71-62032085eb79 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.957076] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3920c9ea-fe6e-4bd8-a1c4-50fd0e804146 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.989026] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a20472a-960e-459b-8796-450335fbd2cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.996484] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3392a19-b301-471b-a939-e128b232bcce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.012025] env[61985]: DEBUG nova.compute.provider_tree [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.383136] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Releasing lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.383136] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 730.383136] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 730.383136] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46418209-e6ac-42a4-97df-1391ca6f350e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.394970] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7096958-f023-4206-8771-a55b447f9ac1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.418104] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 983c68ad-52ca-4eba-9570-3a37d8f5c637 could not be found. [ 730.418407] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 730.418542] env[61985]: INFO nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Took 0.04 seconds to destroy the instance on the hypervisor. [ 730.418857] env[61985]: DEBUG oslo.service.loopingcall [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.419069] env[61985]: DEBUG nova.compute.manager [-] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.419220] env[61985]: DEBUG nova.network.neutron [-] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 730.450064] env[61985]: DEBUG nova.network.neutron [-] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.517437] env[61985]: DEBUG nova.scheduler.client.report [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.565692] env[61985]: DEBUG nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 730.594122] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.594122] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.594122] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.594122] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.594294] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.594294] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.594294] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.594294] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.594294] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.594955] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.595317] env[61985]: DEBUG nova.virt.hardware [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.597065] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81094ffc-9cc3-4346-8d65-3403839f2280 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.606740] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecfe2912-1d46-47eb-bb41-1eda4e5b5f7e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.952669] env[61985]: DEBUG nova.network.neutron [-] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.023026] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.024276] env[61985]: DEBUG nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 731.028248] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.311s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.033702] env[61985]: INFO nova.compute.claims [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.044990] env[61985]: ERROR nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb, please check neutron logs for more information. [ 731.044990] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 731.044990] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.044990] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 731.044990] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 731.044990] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 731.044990] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 731.044990] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 731.044990] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.044990] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 731.044990] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.044990] env[61985]: ERROR nova.compute.manager raise self.value [ 731.044990] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 731.044990] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 731.044990] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.044990] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 731.045418] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.045418] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 731.045418] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb, please check neutron logs for more information. [ 731.045418] env[61985]: ERROR nova.compute.manager [ 731.045418] env[61985]: Traceback (most recent call last): [ 731.045418] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 731.045418] env[61985]: listener.cb(fileno) [ 731.045418] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.045418] env[61985]: result = function(*args, **kwargs) [ 731.045418] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.045418] env[61985]: return func(*args, **kwargs) [ 731.045418] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.045418] env[61985]: raise e [ 731.045418] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.045418] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 731.045418] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 731.045418] env[61985]: created_port_ids = self._update_ports_for_instance( [ 731.045418] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 731.045418] env[61985]: with excutils.save_and_reraise_exception(): [ 731.045418] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.045418] env[61985]: self.force_reraise() [ 731.045418] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.045418] env[61985]: raise self.value [ 731.045418] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 731.045418] env[61985]: updated_port = self._update_port( [ 731.045418] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.045418] env[61985]: _ensure_no_port_binding_failure(port) [ 731.045418] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.045418] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 731.046152] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb, please check neutron logs for more information. [ 731.046152] env[61985]: Removing descriptor: 17 [ 731.046152] env[61985]: ERROR nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb, please check neutron logs for more information. [ 731.046152] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Traceback (most recent call last): [ 731.046152] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 731.046152] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] yield resources [ 731.046152] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 731.046152] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] self.driver.spawn(context, instance, image_meta, [ 731.046152] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 731.046152] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.046152] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.046152] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] vm_ref = self.build_virtual_machine(instance, [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] for vif in network_info: [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] return self._sync_wrapper(fn, *args, **kwargs) [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] self.wait() [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] self[:] = self._gt.wait() [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] return self._exit_event.wait() [ 731.046444] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] result = hub.switch() [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] return self.greenlet.switch() [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] result = function(*args, **kwargs) [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] return func(*args, **kwargs) [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] raise e [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] nwinfo = self.network_api.allocate_for_instance( [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 731.046749] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] created_port_ids = self._update_ports_for_instance( [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] with excutils.save_and_reraise_exception(): [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] self.force_reraise() [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] raise self.value [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] updated_port = self._update_port( [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] _ensure_no_port_binding_failure(port) [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.047070] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] raise exception.PortBindingFailed(port_id=port['id']) [ 731.047765] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] nova.exception.PortBindingFailed: Binding failed for port 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb, please check neutron logs for more information. [ 731.047765] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] [ 731.047765] env[61985]: INFO nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Terminating instance [ 731.048698] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "refresh_cache-c8046008-3645-4845-be1e-23f5fefd5dc6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.049233] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquired lock "refresh_cache-c8046008-3645-4845-be1e-23f5fefd5dc6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.049423] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 731.456463] env[61985]: INFO nova.compute.manager [-] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Took 1.04 seconds to deallocate network for instance. [ 731.458919] env[61985]: DEBUG nova.compute.claims [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 731.459109] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.536743] env[61985]: DEBUG nova.compute.utils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.541319] env[61985]: DEBUG nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 731.544744] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 731.577127] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.647797] env[61985]: DEBUG nova.policy [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb36d8bd7eb64e94b3249f4807c82354', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0fab833015ee44a99955cf9cc959b66a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 731.783669] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.972548] env[61985]: DEBUG nova.compute.manager [req-ef1a3d38-103f-455f-ade0-e0195b037789 req-d603e6be-c80e-4b14-92f1-f48ce0fafe3c service nova] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Received event network-changed-7a4cbcc9-6d36-409d-9f03-9b4a86a456bb {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 731.972548] env[61985]: DEBUG nova.compute.manager [req-ef1a3d38-103f-455f-ade0-e0195b037789 req-d603e6be-c80e-4b14-92f1-f48ce0fafe3c service nova] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Refreshing instance network info cache due to event network-changed-7a4cbcc9-6d36-409d-9f03-9b4a86a456bb. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 731.972692] env[61985]: DEBUG oslo_concurrency.lockutils [req-ef1a3d38-103f-455f-ade0-e0195b037789 req-d603e6be-c80e-4b14-92f1-f48ce0fafe3c service nova] Acquiring lock "refresh_cache-c8046008-3645-4845-be1e-23f5fefd5dc6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.045259] env[61985]: DEBUG nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 732.188670] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Successfully created port: 6812cb3e-1080-408e-8df0-02f671e82b2f {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.237201] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "95e3e6d7-9203-47a0-bad7-050eba09c511" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.237201] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "95e3e6d7-9203-47a0-bad7-050eba09c511" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.274046] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.274342] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.286124] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Releasing lock "refresh_cache-c8046008-3645-4845-be1e-23f5fefd5dc6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.286538] env[61985]: DEBUG nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 732.286690] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 732.287282] env[61985]: DEBUG oslo_concurrency.lockutils [req-ef1a3d38-103f-455f-ade0-e0195b037789 req-d603e6be-c80e-4b14-92f1-f48ce0fafe3c service nova] Acquired lock "refresh_cache-c8046008-3645-4845-be1e-23f5fefd5dc6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.287499] env[61985]: DEBUG nova.network.neutron [req-ef1a3d38-103f-455f-ade0-e0195b037789 req-d603e6be-c80e-4b14-92f1-f48ce0fafe3c service nova] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Refreshing network info cache for port 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 732.288448] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-613d2ef0-dd70-47a5-9257-67bcffb7bcc8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.298088] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92f0625-62a4-4019-abc6-af15421206e2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.324824] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c8046008-3645-4845-be1e-23f5fefd5dc6 could not be found. [ 732.324824] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 732.324824] env[61985]: INFO nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 732.324824] env[61985]: DEBUG oslo.service.loopingcall [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.326706] env[61985]: DEBUG nova.compute.manager [-] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.326806] env[61985]: DEBUG nova.network.neutron [-] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 732.362992] env[61985]: DEBUG nova.network.neutron [-] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 732.555992] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06e1d86-05fc-4787-bb01-a9611e505c8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.563935] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4fc111-4cb6-4e26-b291-17bf26d877fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.594123] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8294e038-9bfc-4fda-b79e-ba7d62f41f0d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.602136] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b120765-9908-47c3-b9df-0bf243c6c506 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.617222] env[61985]: DEBUG nova.compute.provider_tree [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.816857] env[61985]: DEBUG nova.network.neutron [req-ef1a3d38-103f-455f-ade0-e0195b037789 req-d603e6be-c80e-4b14-92f1-f48ce0fafe3c service nova] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 732.865222] env[61985]: DEBUG nova.network.neutron [-] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.961098] env[61985]: DEBUG nova.network.neutron [req-ef1a3d38-103f-455f-ade0-e0195b037789 req-d603e6be-c80e-4b14-92f1-f48ce0fafe3c service nova] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.062642] env[61985]: DEBUG nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 733.090345] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 733.090798] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 733.091894] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.091894] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 733.091894] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.091894] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 733.091894] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 733.092230] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 733.092500] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 733.092777] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 733.093060] env[61985]: DEBUG nova.virt.hardware [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.093991] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a132c40-0c4e-442a-8fc9-31a4cb19afbd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.103981] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7047bd7a-425f-420b-95ce-7123e0a6c7f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.121262] env[61985]: DEBUG nova.scheduler.client.report [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.370446] env[61985]: INFO nova.compute.manager [-] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Took 1.04 seconds to deallocate network for instance. [ 733.371258] env[61985]: DEBUG nova.compute.claims [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 733.371645] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.463766] env[61985]: DEBUG oslo_concurrency.lockutils [req-ef1a3d38-103f-455f-ade0-e0195b037789 req-d603e6be-c80e-4b14-92f1-f48ce0fafe3c service nova] Releasing lock "refresh_cache-c8046008-3645-4845-be1e-23f5fefd5dc6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.464404] env[61985]: DEBUG nova.compute.manager [req-ef1a3d38-103f-455f-ade0-e0195b037789 req-d603e6be-c80e-4b14-92f1-f48ce0fafe3c service nova] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Received event network-vif-deleted-7a4cbcc9-6d36-409d-9f03-9b4a86a456bb {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 733.491172] env[61985]: ERROR nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6812cb3e-1080-408e-8df0-02f671e82b2f, please check neutron logs for more information. [ 733.491172] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 733.491172] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.491172] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 733.491172] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.491172] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 733.491172] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.491172] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 733.491172] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.491172] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 733.491172] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.491172] env[61985]: ERROR nova.compute.manager raise self.value [ 733.491172] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.491172] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 733.491172] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.491172] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 733.491605] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.491605] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 733.491605] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6812cb3e-1080-408e-8df0-02f671e82b2f, please check neutron logs for more information. [ 733.491605] env[61985]: ERROR nova.compute.manager [ 733.491605] env[61985]: Traceback (most recent call last): [ 733.491605] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 733.491605] env[61985]: listener.cb(fileno) [ 733.491605] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.491605] env[61985]: result = function(*args, **kwargs) [ 733.491605] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.491605] env[61985]: return func(*args, **kwargs) [ 733.491605] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.491605] env[61985]: raise e [ 733.491605] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.491605] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 733.491605] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.491605] env[61985]: created_port_ids = self._update_ports_for_instance( [ 733.491605] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.491605] env[61985]: with excutils.save_and_reraise_exception(): [ 733.491605] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.491605] env[61985]: self.force_reraise() [ 733.491605] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.491605] env[61985]: raise self.value [ 733.491605] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.491605] env[61985]: updated_port = self._update_port( [ 733.491605] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.491605] env[61985]: _ensure_no_port_binding_failure(port) [ 733.491605] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.491605] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 733.492299] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 6812cb3e-1080-408e-8df0-02f671e82b2f, please check neutron logs for more information. [ 733.492299] env[61985]: Removing descriptor: 17 [ 733.492299] env[61985]: ERROR nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6812cb3e-1080-408e-8df0-02f671e82b2f, please check neutron logs for more information. [ 733.492299] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Traceback (most recent call last): [ 733.492299] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 733.492299] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] yield resources [ 733.492299] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 733.492299] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] self.driver.spawn(context, instance, image_meta, [ 733.492299] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 733.492299] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.492299] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.492299] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] vm_ref = self.build_virtual_machine(instance, [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] for vif in network_info: [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] return self._sync_wrapper(fn, *args, **kwargs) [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] self.wait() [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] self[:] = self._gt.wait() [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] return self._exit_event.wait() [ 733.492609] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] result = hub.switch() [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] return self.greenlet.switch() [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] result = function(*args, **kwargs) [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] return func(*args, **kwargs) [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] raise e [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] nwinfo = self.network_api.allocate_for_instance( [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.492923] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] created_port_ids = self._update_ports_for_instance( [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] with excutils.save_and_reraise_exception(): [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] self.force_reraise() [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] raise self.value [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] updated_port = self._update_port( [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] _ensure_no_port_binding_failure(port) [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.493239] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] raise exception.PortBindingFailed(port_id=port['id']) [ 733.493527] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] nova.exception.PortBindingFailed: Binding failed for port 6812cb3e-1080-408e-8df0-02f671e82b2f, please check neutron logs for more information. [ 733.493527] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] [ 733.493527] env[61985]: INFO nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Terminating instance [ 733.494509] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Acquiring lock "refresh_cache-3480f0ec-85f4-4881-ad17-d647cd401fbf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.494676] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Acquired lock "refresh_cache-3480f0ec-85f4-4881-ad17-d647cd401fbf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.494845] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 733.627062] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.598s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.627385] env[61985]: DEBUG nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 733.629938] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.151s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.631705] env[61985]: INFO nova.compute.claims [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 734.001083] env[61985]: DEBUG nova.compute.manager [req-854c3c65-b3fe-4221-8d12-901ab08bba37 req-42f2e79e-ff62-41cf-8cc7-84a1a01fe615 service nova] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Received event network-changed-6812cb3e-1080-408e-8df0-02f671e82b2f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 734.001294] env[61985]: DEBUG nova.compute.manager [req-854c3c65-b3fe-4221-8d12-901ab08bba37 req-42f2e79e-ff62-41cf-8cc7-84a1a01fe615 service nova] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Refreshing instance network info cache due to event network-changed-6812cb3e-1080-408e-8df0-02f671e82b2f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 734.001484] env[61985]: DEBUG oslo_concurrency.lockutils [req-854c3c65-b3fe-4221-8d12-901ab08bba37 req-42f2e79e-ff62-41cf-8cc7-84a1a01fe615 service nova] Acquiring lock "refresh_cache-3480f0ec-85f4-4881-ad17-d647cd401fbf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.015194] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.125835] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.135476] env[61985]: DEBUG nova.compute.utils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.139318] env[61985]: DEBUG nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 734.139416] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 734.177594] env[61985]: DEBUG nova.policy [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c75540ff84fa47a4af125ae2ca309597', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83bffe360b4a446ab7705ebe50935307', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.629097] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Releasing lock "refresh_cache-3480f0ec-85f4-4881-ad17-d647cd401fbf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.629097] env[61985]: DEBUG nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 734.630532] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 734.630585] env[61985]: DEBUG oslo_concurrency.lockutils [req-854c3c65-b3fe-4221-8d12-901ab08bba37 req-42f2e79e-ff62-41cf-8cc7-84a1a01fe615 service nova] Acquired lock "refresh_cache-3480f0ec-85f4-4881-ad17-d647cd401fbf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.633527] env[61985]: DEBUG nova.network.neutron [req-854c3c65-b3fe-4221-8d12-901ab08bba37 req-42f2e79e-ff62-41cf-8cc7-84a1a01fe615 service nova] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Refreshing network info cache for port 6812cb3e-1080-408e-8df0-02f671e82b2f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 734.633527] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6df51c48-d4d2-4763-bf89-490a7894c4cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.638872] env[61985]: DEBUG nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 734.649345] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71ba077-0901-456a-9020-97c457b46497 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.676297] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3480f0ec-85f4-4881-ad17-d647cd401fbf could not be found. [ 734.676543] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 734.676729] env[61985]: INFO nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Took 0.05 seconds to destroy the instance on the hypervisor. [ 734.676981] env[61985]: DEBUG oslo.service.loopingcall [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.677224] env[61985]: DEBUG nova.compute.manager [-] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.677383] env[61985]: DEBUG nova.network.neutron [-] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 734.699729] env[61985]: DEBUG nova.network.neutron [-] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.744719] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Successfully created port: 2a7da93d-f288-4c84-98e1-fc3a5403cb1c {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 735.078020] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec9a7df-53a3-4d0e-8543-2484a759b8c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.084564] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad26414-7567-49d5-80cc-b94141a75230 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.128017] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6899646-f892-4f8b-b209-5e7086d46bc6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.134246] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9f4108-47a2-444a-94e8-176317a706da {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.155412] env[61985]: DEBUG nova.compute.provider_tree [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.174490] env[61985]: DEBUG nova.network.neutron [req-854c3c65-b3fe-4221-8d12-901ab08bba37 req-42f2e79e-ff62-41cf-8cc7-84a1a01fe615 service nova] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.202180] env[61985]: DEBUG nova.network.neutron [-] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.302292] env[61985]: DEBUG nova.network.neutron [req-854c3c65-b3fe-4221-8d12-901ab08bba37 req-42f2e79e-ff62-41cf-8cc7-84a1a01fe615 service nova] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.658931] env[61985]: DEBUG nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 735.661732] env[61985]: DEBUG nova.scheduler.client.report [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.686202] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.686453] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.686608] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.686789] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.686945] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.687139] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.687391] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.687602] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.687696] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.687858] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.688033] env[61985]: DEBUG nova.virt.hardware [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.689158] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7a3be5-033a-437f-8ce2-1dc99a977e58 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.697923] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8192ab9-e99a-4d12-aaba-ff5e369b1cc8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.711633] env[61985]: INFO nova.compute.manager [-] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Took 1.03 seconds to deallocate network for instance. [ 735.714118] env[61985]: DEBUG nova.compute.claims [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 735.714308] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.805085] env[61985]: DEBUG oslo_concurrency.lockutils [req-854c3c65-b3fe-4221-8d12-901ab08bba37 req-42f2e79e-ff62-41cf-8cc7-84a1a01fe615 service nova] Releasing lock "refresh_cache-3480f0ec-85f4-4881-ad17-d647cd401fbf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.805313] env[61985]: DEBUG nova.compute.manager [req-854c3c65-b3fe-4221-8d12-901ab08bba37 req-42f2e79e-ff62-41cf-8cc7-84a1a01fe615 service nova] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Received event network-vif-deleted-6812cb3e-1080-408e-8df0-02f671e82b2f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 735.952030] env[61985]: ERROR nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2a7da93d-f288-4c84-98e1-fc3a5403cb1c, please check neutron logs for more information. [ 735.952030] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 735.952030] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.952030] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 735.952030] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.952030] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 735.952030] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.952030] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 735.952030] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.952030] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 735.952030] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.952030] env[61985]: ERROR nova.compute.manager raise self.value [ 735.952030] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.952030] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 735.952030] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.952030] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 735.952747] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.952747] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 735.952747] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2a7da93d-f288-4c84-98e1-fc3a5403cb1c, please check neutron logs for more information. [ 735.952747] env[61985]: ERROR nova.compute.manager [ 735.952747] env[61985]: Traceback (most recent call last): [ 735.952747] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 735.952747] env[61985]: listener.cb(fileno) [ 735.952747] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.952747] env[61985]: result = function(*args, **kwargs) [ 735.952747] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.952747] env[61985]: return func(*args, **kwargs) [ 735.952747] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.952747] env[61985]: raise e [ 735.952747] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.952747] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 735.952747] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.952747] env[61985]: created_port_ids = self._update_ports_for_instance( [ 735.952747] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.952747] env[61985]: with excutils.save_and_reraise_exception(): [ 735.952747] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.952747] env[61985]: self.force_reraise() [ 735.952747] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.952747] env[61985]: raise self.value [ 735.952747] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.952747] env[61985]: updated_port = self._update_port( [ 735.952747] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.952747] env[61985]: _ensure_no_port_binding_failure(port) [ 735.952747] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.952747] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.954075] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 2a7da93d-f288-4c84-98e1-fc3a5403cb1c, please check neutron logs for more information. [ 735.954075] env[61985]: Removing descriptor: 17 [ 735.954075] env[61985]: ERROR nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2a7da93d-f288-4c84-98e1-fc3a5403cb1c, please check neutron logs for more information. [ 735.954075] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Traceback (most recent call last): [ 735.954075] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 735.954075] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] yield resources [ 735.954075] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 735.954075] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] self.driver.spawn(context, instance, image_meta, [ 735.954075] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 735.954075] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.954075] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.954075] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] vm_ref = self.build_virtual_machine(instance, [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] for vif in network_info: [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] return self._sync_wrapper(fn, *args, **kwargs) [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] self.wait() [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] self[:] = self._gt.wait() [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] return self._exit_event.wait() [ 735.954480] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] result = hub.switch() [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] return self.greenlet.switch() [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] result = function(*args, **kwargs) [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] return func(*args, **kwargs) [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] raise e [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] nwinfo = self.network_api.allocate_for_instance( [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.954885] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] created_port_ids = self._update_ports_for_instance( [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] with excutils.save_and_reraise_exception(): [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] self.force_reraise() [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] raise self.value [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] updated_port = self._update_port( [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] _ensure_no_port_binding_failure(port) [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.955278] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] raise exception.PortBindingFailed(port_id=port['id']) [ 735.955624] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] nova.exception.PortBindingFailed: Binding failed for port 2a7da93d-f288-4c84-98e1-fc3a5403cb1c, please check neutron logs for more information. [ 735.955624] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] [ 735.955624] env[61985]: INFO nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Terminating instance [ 735.955624] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Acquiring lock "refresh_cache-4a5bd6dd-9d21-43a2-9108-58aed0637ea9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.955624] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Acquired lock "refresh_cache-4a5bd6dd-9d21-43a2-9108-58aed0637ea9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.955624] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.026030] env[61985]: DEBUG nova.compute.manager [req-30598a33-48c8-4a72-b9e0-aa2a04a77d7c req-07f6ad45-0b13-42bc-8d4f-c34b1c96bce8 service nova] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Received event network-changed-2a7da93d-f288-4c84-98e1-fc3a5403cb1c {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 736.026261] env[61985]: DEBUG nova.compute.manager [req-30598a33-48c8-4a72-b9e0-aa2a04a77d7c req-07f6ad45-0b13-42bc-8d4f-c34b1c96bce8 service nova] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Refreshing instance network info cache due to event network-changed-2a7da93d-f288-4c84-98e1-fc3a5403cb1c. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 736.026455] env[61985]: DEBUG oslo_concurrency.lockutils [req-30598a33-48c8-4a72-b9e0-aa2a04a77d7c req-07f6ad45-0b13-42bc-8d4f-c34b1c96bce8 service nova] Acquiring lock "refresh_cache-4a5bd6dd-9d21-43a2-9108-58aed0637ea9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.166566] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.167085] env[61985]: DEBUG nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 736.170133] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.719s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.171502] env[61985]: INFO nova.compute.claims [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.477103] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.600882] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.677822] env[61985]: DEBUG nova.compute.utils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.679870] env[61985]: DEBUG nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.680058] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 736.735813] env[61985]: DEBUG nova.policy [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c660f9ffa85d4314bcd4051a74b17d4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'efcf3bc5f6d34456aa8b15328208b67f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 737.074931] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Successfully created port: 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.106166] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Releasing lock "refresh_cache-4a5bd6dd-9d21-43a2-9108-58aed0637ea9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.106166] env[61985]: DEBUG nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 737.106166] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 737.106166] env[61985]: DEBUG oslo_concurrency.lockutils [req-30598a33-48c8-4a72-b9e0-aa2a04a77d7c req-07f6ad45-0b13-42bc-8d4f-c34b1c96bce8 service nova] Acquired lock "refresh_cache-4a5bd6dd-9d21-43a2-9108-58aed0637ea9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.106166] env[61985]: DEBUG nova.network.neutron [req-30598a33-48c8-4a72-b9e0-aa2a04a77d7c req-07f6ad45-0b13-42bc-8d4f-c34b1c96bce8 service nova] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Refreshing network info cache for port 2a7da93d-f288-4c84-98e1-fc3a5403cb1c {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 737.106326] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-698f0c09-20bd-4104-93a9-887c1eb9ebc3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.121968] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eedc917-2787-403b-b983-7808271e78c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.155110] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4a5bd6dd-9d21-43a2-9108-58aed0637ea9 could not be found. [ 737.155110] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 737.155110] env[61985]: INFO nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 737.155110] env[61985]: DEBUG oslo.service.loopingcall [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.155110] env[61985]: DEBUG nova.compute.manager [-] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.155110] env[61985]: DEBUG nova.network.neutron [-] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 737.175796] env[61985]: DEBUG nova.network.neutron [-] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.182852] env[61985]: DEBUG nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 737.576295] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd637d91-e970-403b-bdd6-d0620f30dee0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.584070] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1dd1c1-a881-468b-8f43-4ef9a6dec0ed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.614935] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829cb449-eeb6-4181-a939-8624645e026b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.622195] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b8fe75-c610-4703-8773-62424896abe3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.641748] env[61985]: DEBUG nova.compute.provider_tree [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.643544] env[61985]: DEBUG nova.network.neutron [req-30598a33-48c8-4a72-b9e0-aa2a04a77d7c req-07f6ad45-0b13-42bc-8d4f-c34b1c96bce8 service nova] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.678479] env[61985]: DEBUG nova.network.neutron [-] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.858209] env[61985]: DEBUG nova.network.neutron [req-30598a33-48c8-4a72-b9e0-aa2a04a77d7c req-07f6ad45-0b13-42bc-8d4f-c34b1c96bce8 service nova] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.146989] env[61985]: DEBUG nova.scheduler.client.report [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.181855] env[61985]: INFO nova.compute.manager [-] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Took 1.03 seconds to deallocate network for instance. [ 738.183421] env[61985]: DEBUG nova.compute.claims [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 738.183868] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.195071] env[61985]: DEBUG nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 738.223284] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.223539] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.223696] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.223876] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.224039] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.224193] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.224401] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.224562] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.224728] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.224889] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.225406] env[61985]: DEBUG nova.virt.hardware [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.225955] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddc5f1c-7067-4cd0-a7b9-757bfa7e9ed0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.230623] env[61985]: DEBUG nova.compute.manager [req-736f0199-32a2-4d07-9c31-cc47ecc57941 req-f4f195b3-004c-4edf-a949-4ee8324a39da service nova] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Received event network-changed-8a26173e-d5a2-4f1a-bd7d-3f032f0067cc {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 738.230804] env[61985]: DEBUG nova.compute.manager [req-736f0199-32a2-4d07-9c31-cc47ecc57941 req-f4f195b3-004c-4edf-a949-4ee8324a39da service nova] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Refreshing instance network info cache due to event network-changed-8a26173e-d5a2-4f1a-bd7d-3f032f0067cc. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 738.231016] env[61985]: DEBUG oslo_concurrency.lockutils [req-736f0199-32a2-4d07-9c31-cc47ecc57941 req-f4f195b3-004c-4edf-a949-4ee8324a39da service nova] Acquiring lock "refresh_cache-9af10e29-12ea-441d-b82a-be6976032c88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.231164] env[61985]: DEBUG oslo_concurrency.lockutils [req-736f0199-32a2-4d07-9c31-cc47ecc57941 req-f4f195b3-004c-4edf-a949-4ee8324a39da service nova] Acquired lock "refresh_cache-9af10e29-12ea-441d-b82a-be6976032c88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.231321] env[61985]: DEBUG nova.network.neutron [req-736f0199-32a2-4d07-9c31-cc47ecc57941 req-f4f195b3-004c-4edf-a949-4ee8324a39da service nova] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Refreshing network info cache for port 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 738.238249] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b2f168-6750-4d9a-8a07-e0ce120aa515 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.362668] env[61985]: DEBUG oslo_concurrency.lockutils [req-30598a33-48c8-4a72-b9e0-aa2a04a77d7c req-07f6ad45-0b13-42bc-8d4f-c34b1c96bce8 service nova] Releasing lock "refresh_cache-4a5bd6dd-9d21-43a2-9108-58aed0637ea9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.362668] env[61985]: DEBUG nova.compute.manager [req-30598a33-48c8-4a72-b9e0-aa2a04a77d7c req-07f6ad45-0b13-42bc-8d4f-c34b1c96bce8 service nova] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Received event network-vif-deleted-2a7da93d-f288-4c84-98e1-fc3a5403cb1c {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 738.400125] env[61985]: ERROR nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc, please check neutron logs for more information. [ 738.400125] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 738.400125] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.400125] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 738.400125] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.400125] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 738.400125] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.400125] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 738.400125] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.400125] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 738.400125] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.400125] env[61985]: ERROR nova.compute.manager raise self.value [ 738.400125] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.400125] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 738.400125] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.400125] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 738.400619] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.400619] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 738.400619] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc, please check neutron logs for more information. [ 738.400619] env[61985]: ERROR nova.compute.manager [ 738.400619] env[61985]: Traceback (most recent call last): [ 738.400619] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 738.400619] env[61985]: listener.cb(fileno) [ 738.400619] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.400619] env[61985]: result = function(*args, **kwargs) [ 738.400619] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.400619] env[61985]: return func(*args, **kwargs) [ 738.400619] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.400619] env[61985]: raise e [ 738.400619] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.400619] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 738.400619] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.400619] env[61985]: created_port_ids = self._update_ports_for_instance( [ 738.400619] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.400619] env[61985]: with excutils.save_and_reraise_exception(): [ 738.400619] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.400619] env[61985]: self.force_reraise() [ 738.400619] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.400619] env[61985]: raise self.value [ 738.400619] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.400619] env[61985]: updated_port = self._update_port( [ 738.400619] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.400619] env[61985]: _ensure_no_port_binding_failure(port) [ 738.400619] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.400619] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 738.401453] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc, please check neutron logs for more information. [ 738.401453] env[61985]: Removing descriptor: 17 [ 738.401453] env[61985]: ERROR nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc, please check neutron logs for more information. [ 738.401453] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Traceback (most recent call last): [ 738.401453] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 738.401453] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] yield resources [ 738.401453] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 738.401453] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] self.driver.spawn(context, instance, image_meta, [ 738.401453] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 738.401453] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.401453] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.401453] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] vm_ref = self.build_virtual_machine(instance, [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] for vif in network_info: [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] return self._sync_wrapper(fn, *args, **kwargs) [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] self.wait() [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] self[:] = self._gt.wait() [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] return self._exit_event.wait() [ 738.401810] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] result = hub.switch() [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] return self.greenlet.switch() [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] result = function(*args, **kwargs) [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] return func(*args, **kwargs) [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] raise e [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] nwinfo = self.network_api.allocate_for_instance( [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.402158] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] created_port_ids = self._update_ports_for_instance( [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] with excutils.save_and_reraise_exception(): [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] self.force_reraise() [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] raise self.value [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] updated_port = self._update_port( [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] _ensure_no_port_binding_failure(port) [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.402485] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] raise exception.PortBindingFailed(port_id=port['id']) [ 738.402774] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] nova.exception.PortBindingFailed: Binding failed for port 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc, please check neutron logs for more information. [ 738.402774] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] [ 738.402774] env[61985]: INFO nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Terminating instance [ 738.406506] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Acquiring lock "refresh_cache-9af10e29-12ea-441d-b82a-be6976032c88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.652529] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.653079] env[61985]: DEBUG nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 738.655679] env[61985]: DEBUG oslo_concurrency.lockutils [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 33.905s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.655861] env[61985]: DEBUG nova.objects.instance [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61985) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 738.830564] env[61985]: DEBUG nova.network.neutron [req-736f0199-32a2-4d07-9c31-cc47ecc57941 req-f4f195b3-004c-4edf-a949-4ee8324a39da service nova] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.948400] env[61985]: DEBUG nova.network.neutron [req-736f0199-32a2-4d07-9c31-cc47ecc57941 req-f4f195b3-004c-4edf-a949-4ee8324a39da service nova] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.160470] env[61985]: DEBUG nova.compute.utils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 739.164544] env[61985]: DEBUG nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 739.164719] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 739.224327] env[61985]: DEBUG nova.policy [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '610d4416c1304ac29ef1889d5a227a77', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9e1a2056faf4d0fa18aaec978e8f374', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.450666] env[61985]: DEBUG oslo_concurrency.lockutils [req-736f0199-32a2-4d07-9c31-cc47ecc57941 req-f4f195b3-004c-4edf-a949-4ee8324a39da service nova] Releasing lock "refresh_cache-9af10e29-12ea-441d-b82a-be6976032c88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.451096] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Acquired lock "refresh_cache-9af10e29-12ea-441d-b82a-be6976032c88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.451291] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 739.570240] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Successfully created port: 10db0b21-b9cc-4931-b892-72fc107d3856 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.665752] env[61985]: DEBUG nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 739.669126] env[61985]: DEBUG oslo_concurrency.lockutils [None req-795f8a60-9622-4d69-b2b1-97811c236ddf tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.670191] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.626s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.670410] env[61985]: DEBUG nova.objects.instance [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lazy-loading 'resources' on Instance uuid 2e1e46ea-477c-41f3-938f-bd654bc5999d {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 739.977834] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.154776] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.294141] env[61985]: DEBUG nova.compute.manager [req-b9bf5a41-ab07-4e17-8060-71535aa2e6c0 req-bdfe83de-4d6f-4eca-a1b3-767f8743267f service nova] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Received event network-vif-deleted-8a26173e-d5a2-4f1a-bd7d-3f032f0067cc {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 740.630421] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0350424-5777-48ce-bbd9-a47acb85109d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.638923] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b11c089-d5db-402f-99ab-0f0119f30a67 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.668193] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Releasing lock "refresh_cache-9af10e29-12ea-441d-b82a-be6976032c88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.668635] env[61985]: DEBUG nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 740.668845] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 740.669356] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-adf31c62-022d-4d3d-8969-3490495bcb38 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.671621] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17144de0-f0b0-4ac6-b4de-8d1ed2017fc4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.679755] env[61985]: DEBUG nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 740.685854] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6919bd-61bc-4c19-9c2a-b908cb59f93a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.696235] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e130b9-872d-4857-a210-395d60886ed0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.719531] env[61985]: DEBUG nova.compute.provider_tree [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.726318] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.726546] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.726699] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.726873] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.727022] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.729446] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.729446] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.729446] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.729446] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.729446] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.729808] env[61985]: DEBUG nova.virt.hardware [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.729808] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9af10e29-12ea-441d-b82a-be6976032c88 could not be found. [ 740.729808] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 740.729808] env[61985]: INFO nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Took 0.06 seconds to destroy the instance on the hypervisor. [ 740.729808] env[61985]: DEBUG oslo.service.loopingcall [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.730063] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab3f357-3ad8-42fd-b5e0-a69fe07d1684 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.733243] env[61985]: DEBUG nova.compute.manager [-] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.733341] env[61985]: DEBUG nova.network.neutron [-] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 740.740405] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af1aa02-d24f-4eb5-bbe8-edc663931b79 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.761403] env[61985]: DEBUG nova.network.neutron [-] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.884250] env[61985]: ERROR nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 10db0b21-b9cc-4931-b892-72fc107d3856, please check neutron logs for more information. [ 740.884250] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 740.884250] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.884250] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 740.884250] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 740.884250] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 740.884250] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 740.884250] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 740.884250] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.884250] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 740.884250] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.884250] env[61985]: ERROR nova.compute.manager raise self.value [ 740.884250] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 740.884250] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 740.884250] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.884250] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 740.884782] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.884782] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 740.884782] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 10db0b21-b9cc-4931-b892-72fc107d3856, please check neutron logs for more information. [ 740.884782] env[61985]: ERROR nova.compute.manager [ 740.884782] env[61985]: Traceback (most recent call last): [ 740.884782] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 740.884782] env[61985]: listener.cb(fileno) [ 740.884782] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.884782] env[61985]: result = function(*args, **kwargs) [ 740.884782] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.884782] env[61985]: return func(*args, **kwargs) [ 740.884782] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.884782] env[61985]: raise e [ 740.884782] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.884782] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 740.884782] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 740.884782] env[61985]: created_port_ids = self._update_ports_for_instance( [ 740.884782] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 740.884782] env[61985]: with excutils.save_and_reraise_exception(): [ 740.884782] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.884782] env[61985]: self.force_reraise() [ 740.884782] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.884782] env[61985]: raise self.value [ 740.884782] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 740.884782] env[61985]: updated_port = self._update_port( [ 740.884782] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.884782] env[61985]: _ensure_no_port_binding_failure(port) [ 740.884782] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.884782] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 740.885695] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 10db0b21-b9cc-4931-b892-72fc107d3856, please check neutron logs for more information. [ 740.885695] env[61985]: Removing descriptor: 17 [ 740.885695] env[61985]: ERROR nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 10db0b21-b9cc-4931-b892-72fc107d3856, please check neutron logs for more information. [ 740.885695] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Traceback (most recent call last): [ 740.885695] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 740.885695] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] yield resources [ 740.885695] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 740.885695] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] self.driver.spawn(context, instance, image_meta, [ 740.885695] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 740.885695] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.885695] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.885695] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] vm_ref = self.build_virtual_machine(instance, [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] for vif in network_info: [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] return self._sync_wrapper(fn, *args, **kwargs) [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] self.wait() [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] self[:] = self._gt.wait() [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] return self._exit_event.wait() [ 740.886099] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] result = hub.switch() [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] return self.greenlet.switch() [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] result = function(*args, **kwargs) [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] return func(*args, **kwargs) [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] raise e [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] nwinfo = self.network_api.allocate_for_instance( [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 740.886492] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] created_port_ids = self._update_ports_for_instance( [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] with excutils.save_and_reraise_exception(): [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] self.force_reraise() [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] raise self.value [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] updated_port = self._update_port( [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] _ensure_no_port_binding_failure(port) [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.886903] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] raise exception.PortBindingFailed(port_id=port['id']) [ 740.887289] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] nova.exception.PortBindingFailed: Binding failed for port 10db0b21-b9cc-4931-b892-72fc107d3856, please check neutron logs for more information. [ 740.887289] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] [ 740.887289] env[61985]: INFO nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Terminating instance [ 740.887427] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Acquiring lock "refresh_cache-a0fd407f-7cf0-4c01-8195-399d8cbf62f9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.887650] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Acquired lock "refresh_cache-a0fd407f-7cf0-4c01-8195-399d8cbf62f9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.887821] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.236180] env[61985]: DEBUG nova.scheduler.client.report [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.264187] env[61985]: DEBUG nova.network.neutron [-] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.423017] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.513660] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.739662] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.069s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.742193] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.868s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.743897] env[61985]: INFO nova.compute.claims [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.761171] env[61985]: INFO nova.scheduler.client.report [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleted allocations for instance 2e1e46ea-477c-41f3-938f-bd654bc5999d [ 741.766280] env[61985]: INFO nova.compute.manager [-] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Took 1.03 seconds to deallocate network for instance. [ 741.770903] env[61985]: DEBUG nova.compute.claims [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 741.770903] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.017020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Releasing lock "refresh_cache-a0fd407f-7cf0-4c01-8195-399d8cbf62f9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.017557] env[61985]: DEBUG nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 742.017867] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 742.018200] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6b2a4f0-4ba8-475c-9f46-065f50792212 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.027300] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad847fa1-5698-4ef6-a709-007e8552371d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.049167] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a0fd407f-7cf0-4c01-8195-399d8cbf62f9 could not be found. [ 742.049351] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 742.049539] env[61985]: INFO nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 742.049805] env[61985]: DEBUG oslo.service.loopingcall [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.050024] env[61985]: DEBUG nova.compute.manager [-] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.050108] env[61985]: DEBUG nova.network.neutron [-] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 742.064090] env[61985]: DEBUG nova.network.neutron [-] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.268029] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c8c0511c-229e-4e8a-8f87-f39e126e2f3b tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "2e1e46ea-477c-41f3-938f-bd654bc5999d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.912s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.339220] env[61985]: DEBUG nova.compute.manager [req-f24c02b3-1d6a-4737-a207-ff4bdcabb71f req-c3e35589-d3fb-46eb-91c2-7c73e0e1df0a service nova] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Received event network-changed-10db0b21-b9cc-4931-b892-72fc107d3856 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 742.339385] env[61985]: DEBUG nova.compute.manager [req-f24c02b3-1d6a-4737-a207-ff4bdcabb71f req-c3e35589-d3fb-46eb-91c2-7c73e0e1df0a service nova] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Refreshing instance network info cache due to event network-changed-10db0b21-b9cc-4931-b892-72fc107d3856. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 742.339608] env[61985]: DEBUG oslo_concurrency.lockutils [req-f24c02b3-1d6a-4737-a207-ff4bdcabb71f req-c3e35589-d3fb-46eb-91c2-7c73e0e1df0a service nova] Acquiring lock "refresh_cache-a0fd407f-7cf0-4c01-8195-399d8cbf62f9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.339894] env[61985]: DEBUG oslo_concurrency.lockutils [req-f24c02b3-1d6a-4737-a207-ff4bdcabb71f req-c3e35589-d3fb-46eb-91c2-7c73e0e1df0a service nova] Acquired lock "refresh_cache-a0fd407f-7cf0-4c01-8195-399d8cbf62f9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.339894] env[61985]: DEBUG nova.network.neutron [req-f24c02b3-1d6a-4737-a207-ff4bdcabb71f req-c3e35589-d3fb-46eb-91c2-7c73e0e1df0a service nova] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Refreshing network info cache for port 10db0b21-b9cc-4931-b892-72fc107d3856 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 742.567131] env[61985]: DEBUG nova.network.neutron [-] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.611023] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "213a3e19-5589-4261-96b0-69acfb6798ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.611117] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "213a3e19-5589-4261-96b0-69acfb6798ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.611318] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "213a3e19-5589-4261-96b0-69acfb6798ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.611502] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "213a3e19-5589-4261-96b0-69acfb6798ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.611672] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "213a3e19-5589-4261-96b0-69acfb6798ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.614150] env[61985]: INFO nova.compute.manager [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Terminating instance [ 742.615805] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "refresh_cache-213a3e19-5589-4261-96b0-69acfb6798ef" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.615964] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquired lock "refresh_cache-213a3e19-5589-4261-96b0-69acfb6798ef" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.616151] env[61985]: DEBUG nova.network.neutron [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.858537] env[61985]: DEBUG nova.network.neutron [req-f24c02b3-1d6a-4737-a207-ff4bdcabb71f req-c3e35589-d3fb-46eb-91c2-7c73e0e1df0a service nova] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.958107] env[61985]: DEBUG nova.network.neutron [req-f24c02b3-1d6a-4737-a207-ff4bdcabb71f req-c3e35589-d3fb-46eb-91c2-7c73e0e1df0a service nova] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.071668] env[61985]: INFO nova.compute.manager [-] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Took 1.02 seconds to deallocate network for instance. [ 743.073860] env[61985]: DEBUG nova.compute.claims [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 743.074053] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.088666] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8d4d0c-5653-4aed-bc2c-7d804bfd8fba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.096796] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb89eba-372a-430c-babc-540dd8db28ec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.131387] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb485196-80f5-4dc4-a85c-68308f1e19aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.138898] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc56161-8108-4ea6-bf9d-5603bd40d3d7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.152743] env[61985]: DEBUG nova.compute.provider_tree [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.154498] env[61985]: DEBUG nova.network.neutron [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.220844] env[61985]: DEBUG nova.network.neutron [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.460838] env[61985]: DEBUG oslo_concurrency.lockutils [req-f24c02b3-1d6a-4737-a207-ff4bdcabb71f req-c3e35589-d3fb-46eb-91c2-7c73e0e1df0a service nova] Releasing lock "refresh_cache-a0fd407f-7cf0-4c01-8195-399d8cbf62f9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.461174] env[61985]: DEBUG nova.compute.manager [req-f24c02b3-1d6a-4737-a207-ff4bdcabb71f req-c3e35589-d3fb-46eb-91c2-7c73e0e1df0a service nova] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Received event network-vif-deleted-10db0b21-b9cc-4931-b892-72fc107d3856 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 743.658533] env[61985]: DEBUG nova.scheduler.client.report [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.725073] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Releasing lock "refresh_cache-213a3e19-5589-4261-96b0-69acfb6798ef" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.725540] env[61985]: DEBUG nova.compute.manager [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 743.725735] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 743.726633] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf10d05-40e0-4259-aef6-66901b12643d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.734825] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 743.735083] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3db2f48-e6b2-496e-ada4-a9a1df497a8f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.743012] env[61985]: DEBUG oslo_vmware.api [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 743.743012] env[61985]: value = "task-935654" [ 743.743012] env[61985]: _type = "Task" [ 743.743012] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.750243] env[61985]: DEBUG oslo_vmware.api [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935654, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.162939] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.163498] env[61985]: DEBUG nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 744.166266] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.159s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.251870] env[61985]: DEBUG oslo_vmware.api [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935654, 'name': PowerOffVM_Task, 'duration_secs': 0.125951} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.252234] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 744.252454] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 744.252750] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4fe4b3a5-c64d-4582-8038-4d9c0e1b7751 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.278438] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 744.278690] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 744.278882] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleting the datastore file [datastore1] 213a3e19-5589-4261-96b0-69acfb6798ef {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 744.279181] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3cea63ae-962a-4b49-a896-626d1b79af62 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.286040] env[61985]: DEBUG oslo_vmware.api [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for the task: (returnval){ [ 744.286040] env[61985]: value = "task-935656" [ 744.286040] env[61985]: _type = "Task" [ 744.286040] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.293380] env[61985]: DEBUG oslo_vmware.api [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.674166] env[61985]: DEBUG nova.compute.utils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.676126] env[61985]: DEBUG nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.676331] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 744.735795] env[61985]: DEBUG nova.policy [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6fc9cde9322d4eefb2472238179f0519', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a609839c3edd4441a3fed8317142fc5e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 744.797932] env[61985]: DEBUG oslo_vmware.api [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Task: {'id': task-935656, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102882} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.799215] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 744.799215] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 744.799215] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 744.799215] env[61985]: INFO nova.compute.manager [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Took 1.07 seconds to destroy the instance on the hypervisor. [ 744.799215] env[61985]: DEBUG oslo.service.loopingcall [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.799475] env[61985]: DEBUG nova.compute.manager [-] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.799508] env[61985]: DEBUG nova.network.neutron [-] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 744.820617] env[61985]: DEBUG nova.network.neutron [-] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.047442] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc53d11-4a2a-412b-a97c-de05acff4009 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.055313] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e061bb2b-fb7e-4e65-a952-d40b833babc9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.061831] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Successfully created port: 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.089547] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9f32a9-7212-447d-9d2f-9414f34e8411 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.098017] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8bb2bdb-120b-4406-b839-96f544f154f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.110377] env[61985]: DEBUG nova.compute.provider_tree [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 745.179189] env[61985]: DEBUG nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 745.323306] env[61985]: DEBUG nova.network.neutron [-] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.632415] env[61985]: ERROR nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [req-6aef6c11-8aac-46f6-b3f9-a4fe11533bd1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6aef6c11-8aac-46f6-b3f9-a4fe11533bd1"}]}: nova.exception.PortBindingFailed: Binding failed for port b51b28a0-1c84-4229-895b-4546f9a0c449, please check neutron logs for more information. [ 745.652155] env[61985]: DEBUG nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 745.664857] env[61985]: DEBUG nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 745.665098] env[61985]: DEBUG nova.compute.provider_tree [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 745.677920] env[61985]: DEBUG nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 745.709309] env[61985]: DEBUG nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 745.825756] env[61985]: INFO nova.compute.manager [-] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Took 1.03 seconds to deallocate network for instance. [ 746.116058] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f047ee7-db29-4054-ad66-43deff955eca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.123065] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c5065a-729a-4c66-9a12-fc0d44e610a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.152977] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff44ec7-f4da-48f5-82b7-707feac57283 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.159713] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5dcc6ed-b0ea-4cfb-abdc-3709c28ec34d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.172738] env[61985]: DEBUG nova.compute.provider_tree [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 746.189362] env[61985]: DEBUG nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 746.222149] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.222403] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.222555] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.222737] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.222885] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.223043] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.223259] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.223423] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.223596] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.223760] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.223932] env[61985]: DEBUG nova.virt.hardware [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.224831] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46351f2e-f5f2-4730-a5ee-bc3e13538da9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.233162] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3878cf45-865f-44c8-8f11-0ab8142f8380 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.279579] env[61985]: DEBUG nova.compute.manager [req-e03151e5-29db-4e11-984d-020f7a21a58d req-774984fd-95bb-4150-acdc-30f3630f12e1 service nova] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Received event network-changed-7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 746.279909] env[61985]: DEBUG nova.compute.manager [req-e03151e5-29db-4e11-984d-020f7a21a58d req-774984fd-95bb-4150-acdc-30f3630f12e1 service nova] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Refreshing instance network info cache due to event network-changed-7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 746.280070] env[61985]: DEBUG oslo_concurrency.lockutils [req-e03151e5-29db-4e11-984d-020f7a21a58d req-774984fd-95bb-4150-acdc-30f3630f12e1 service nova] Acquiring lock "refresh_cache-50315640-7609-4fff-b191-aa29f06cb4f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.280232] env[61985]: DEBUG oslo_concurrency.lockutils [req-e03151e5-29db-4e11-984d-020f7a21a58d req-774984fd-95bb-4150-acdc-30f3630f12e1 service nova] Acquired lock "refresh_cache-50315640-7609-4fff-b191-aa29f06cb4f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.280533] env[61985]: DEBUG nova.network.neutron [req-e03151e5-29db-4e11-984d-020f7a21a58d req-774984fd-95bb-4150-acdc-30f3630f12e1 service nova] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Refreshing network info cache for port 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 746.331655] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.447810] env[61985]: ERROR nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2, please check neutron logs for more information. [ 746.447810] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.447810] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.447810] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.447810] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 746.447810] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.447810] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 746.447810] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.447810] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.447810] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 746.447810] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.447810] env[61985]: ERROR nova.compute.manager raise self.value [ 746.447810] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 746.447810] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.447810] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.447810] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.448328] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.448328] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.448328] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2, please check neutron logs for more information. [ 746.448328] env[61985]: ERROR nova.compute.manager [ 746.448328] env[61985]: Traceback (most recent call last): [ 746.448328] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.448328] env[61985]: listener.cb(fileno) [ 746.448328] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.448328] env[61985]: result = function(*args, **kwargs) [ 746.448328] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.448328] env[61985]: return func(*args, **kwargs) [ 746.448328] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.448328] env[61985]: raise e [ 746.448328] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.448328] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 746.448328] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 746.448328] env[61985]: created_port_ids = self._update_ports_for_instance( [ 746.448328] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 746.448328] env[61985]: with excutils.save_and_reraise_exception(): [ 746.448328] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.448328] env[61985]: self.force_reraise() [ 746.448328] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.448328] env[61985]: raise self.value [ 746.448328] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 746.448328] env[61985]: updated_port = self._update_port( [ 746.448328] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.448328] env[61985]: _ensure_no_port_binding_failure(port) [ 746.448328] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.448328] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.449117] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2, please check neutron logs for more information. [ 746.449117] env[61985]: Removing descriptor: 17 [ 746.449596] env[61985]: ERROR nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2, please check neutron logs for more information. [ 746.449596] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Traceback (most recent call last): [ 746.449596] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 746.449596] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] yield resources [ 746.449596] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 746.449596] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] self.driver.spawn(context, instance, image_meta, [ 746.449596] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 746.449596] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.449596] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.449596] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] vm_ref = self.build_virtual_machine(instance, [ 746.449596] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] for vif in network_info: [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] return self._sync_wrapper(fn, *args, **kwargs) [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] self.wait() [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] self[:] = self._gt.wait() [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] return self._exit_event.wait() [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 746.450050] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] result = hub.switch() [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] return self.greenlet.switch() [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] result = function(*args, **kwargs) [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] return func(*args, **kwargs) [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] raise e [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] nwinfo = self.network_api.allocate_for_instance( [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] created_port_ids = self._update_ports_for_instance( [ 746.450409] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] with excutils.save_and_reraise_exception(): [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] self.force_reraise() [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] raise self.value [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] updated_port = self._update_port( [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] _ensure_no_port_binding_failure(port) [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] raise exception.PortBindingFailed(port_id=port['id']) [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] nova.exception.PortBindingFailed: Binding failed for port 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2, please check neutron logs for more information. [ 746.450755] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] [ 746.451085] env[61985]: INFO nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Terminating instance [ 746.452489] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Acquiring lock "refresh_cache-50315640-7609-4fff-b191-aa29f06cb4f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.707561] env[61985]: DEBUG nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 63 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 746.707925] env[61985]: DEBUG nova.compute.provider_tree [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 63 to 64 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 746.708165] env[61985]: DEBUG nova.compute.provider_tree [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 746.796524] env[61985]: DEBUG nova.network.neutron [req-e03151e5-29db-4e11-984d-020f7a21a58d req-774984fd-95bb-4150-acdc-30f3630f12e1 service nova] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.906079] env[61985]: DEBUG nova.network.neutron [req-e03151e5-29db-4e11-984d-020f7a21a58d req-774984fd-95bb-4150-acdc-30f3630f12e1 service nova] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.213365] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 3.047s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.213944] env[61985]: ERROR nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b51b28a0-1c84-4229-895b-4546f9a0c449, please check neutron logs for more information. [ 747.213944] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Traceback (most recent call last): [ 747.213944] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 747.213944] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] self.driver.spawn(context, instance, image_meta, [ 747.213944] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 747.213944] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.213944] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.213944] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] vm_ref = self.build_virtual_machine(instance, [ 747.213944] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.213944] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.213944] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] for vif in network_info: [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] return self._sync_wrapper(fn, *args, **kwargs) [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] self.wait() [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] self[:] = self._gt.wait() [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] return self._exit_event.wait() [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] result = hub.switch() [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 747.214321] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] return self.greenlet.switch() [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] result = function(*args, **kwargs) [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] return func(*args, **kwargs) [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] raise e [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] nwinfo = self.network_api.allocate_for_instance( [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] created_port_ids = self._update_ports_for_instance( [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] with excutils.save_and_reraise_exception(): [ 747.214739] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] self.force_reraise() [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] raise self.value [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] updated_port = self._update_port( [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] _ensure_no_port_binding_failure(port) [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] raise exception.PortBindingFailed(port_id=port['id']) [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] nova.exception.PortBindingFailed: Binding failed for port b51b28a0-1c84-4229-895b-4546f9a0c449, please check neutron logs for more information. [ 747.215060] env[61985]: ERROR nova.compute.manager [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] [ 747.215359] env[61985]: DEBUG nova.compute.utils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Binding failed for port b51b28a0-1c84-4229-895b-4546f9a0c449, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 747.216288] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.005s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.219116] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Build of instance 40548796-fb8b-4b7e-9bfe-e83ca0edc01f was re-scheduled: Binding failed for port b51b28a0-1c84-4229-895b-4546f9a0c449, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 747.219590] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 747.219816] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "refresh_cache-40548796-fb8b-4b7e-9bfe-e83ca0edc01f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.219964] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquired lock "refresh_cache-40548796-fb8b-4b7e-9bfe-e83ca0edc01f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.220141] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 747.408267] env[61985]: DEBUG oslo_concurrency.lockutils [req-e03151e5-29db-4e11-984d-020f7a21a58d req-774984fd-95bb-4150-acdc-30f3630f12e1 service nova] Releasing lock "refresh_cache-50315640-7609-4fff-b191-aa29f06cb4f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.408744] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Acquired lock "refresh_cache-50315640-7609-4fff-b191-aa29f06cb4f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.408953] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 747.749053] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 747.839510] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.931367] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.017734] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.073244] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0946cf8-eeb1-4159-858b-abad6bc82328 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.080798] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4ef250-91e8-4e70-8e26-aed27e7115c2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.112238] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5bde65-2dd4-4e11-be3d-96908a39c852 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.120105] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed6f4ca-4fb7-486c-b196-d2afcb4059cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.133655] env[61985]: DEBUG nova.compute.provider_tree [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.304775] env[61985]: DEBUG nova.compute.manager [req-6047e62b-a155-4176-95ba-f7386f3f5054 req-354a57e3-1d00-4aaa-a5a8-e7ef7cde03c3 service nova] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Received event network-vif-deleted-7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 748.342629] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Releasing lock "refresh_cache-40548796-fb8b-4b7e-9bfe-e83ca0edc01f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.342903] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 748.343114] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.343286] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 748.365069] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.520026] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Releasing lock "refresh_cache-50315640-7609-4fff-b191-aa29f06cb4f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.520498] env[61985]: DEBUG nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 748.520706] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 748.521009] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ebd8dd0-1818-439e-baeb-5ecd17babeb8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.530515] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9bc5f8b-f4ed-457f-aafc-00b982b4afe2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.553442] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 50315640-7609-4fff-b191-aa29f06cb4f1 could not be found. [ 748.553649] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 748.553837] env[61985]: INFO nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 748.554137] env[61985]: DEBUG oslo.service.loopingcall [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.554343] env[61985]: DEBUG nova.compute.manager [-] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.554404] env[61985]: DEBUG nova.network.neutron [-] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 748.573836] env[61985]: DEBUG nova.network.neutron [-] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.636561] env[61985]: DEBUG nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.871175] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.076026] env[61985]: DEBUG nova.network.neutron [-] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.142175] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.925s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.142175] env[61985]: ERROR nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. [ 749.142175] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Traceback (most recent call last): [ 749.142175] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 749.142175] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] self.driver.spawn(context, instance, image_meta, [ 749.142175] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 749.142175] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.142175] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.142175] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] vm_ref = self.build_virtual_machine(instance, [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] for vif in network_info: [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] return self._sync_wrapper(fn, *args, **kwargs) [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] self.wait() [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] self[:] = self._gt.wait() [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] return self._exit_event.wait() [ 749.142576] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] result = hub.switch() [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] return self.greenlet.switch() [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] result = function(*args, **kwargs) [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] return func(*args, **kwargs) [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] raise e [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] nwinfo = self.network_api.allocate_for_instance( [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.142969] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] created_port_ids = self._update_ports_for_instance( [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] with excutils.save_and_reraise_exception(): [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] self.force_reraise() [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] raise self.value [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] updated_port = self._update_port( [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] _ensure_no_port_binding_failure(port) [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.143383] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] raise exception.PortBindingFailed(port_id=port['id']) [ 749.143759] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] nova.exception.PortBindingFailed: Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. [ 749.143759] env[61985]: ERROR nova.compute.manager [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] [ 749.143759] env[61985]: DEBUG nova.compute.utils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.143861] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.081s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.145962] env[61985]: INFO nova.compute.claims [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.149034] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Build of instance 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d was re-scheduled: Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 749.149234] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 749.149429] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "refresh_cache-6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.149576] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquired lock "refresh_cache-6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.149738] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 749.373241] env[61985]: INFO nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 40548796-fb8b-4b7e-9bfe-e83ca0edc01f] Took 1.03 seconds to deallocate network for instance. [ 749.578794] env[61985]: INFO nova.compute.manager [-] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Took 1.02 seconds to deallocate network for instance. [ 749.581270] env[61985]: DEBUG nova.compute.claims [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 749.581482] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.672034] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 749.785431] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.289878] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Releasing lock "refresh_cache-6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.289878] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 750.289878] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.289878] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 750.305509] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.398594] env[61985]: INFO nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Deleted allocations for instance 40548796-fb8b-4b7e-9bfe-e83ca0edc01f [ 750.512262] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c046336-ed4a-462d-9ebe-89c603d12df2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.520124] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeadc86b-43cd-430e-9053-07fc0b8a0abe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.550735] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587bc05c-b8fd-418b-9b2b-9da17ccdb0fe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.557985] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102a38f9-92e0-44de-a479-6f797786f245 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.570556] env[61985]: DEBUG nova.compute.provider_tree [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.808626] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.908147] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "40548796-fb8b-4b7e-9bfe-e83ca0edc01f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.438s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.073577] env[61985]: DEBUG nova.scheduler.client.report [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.310303] env[61985]: INFO nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d] Took 1.02 seconds to deallocate network for instance. [ 751.411351] env[61985]: DEBUG nova.compute.manager [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 751.581353] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.581353] env[61985]: DEBUG nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 751.583315] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.124s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.930606] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.090628] env[61985]: DEBUG nova.compute.utils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.092087] env[61985]: DEBUG nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 752.092265] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 752.148433] env[61985]: DEBUG nova.policy [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '531a544948ef4e87bd4a9d13fda0683f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd331cf25db143459cd2a50692bc72ed', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 752.274423] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "983c68ad-52ca-4eba-9570-3a37d8f5c637" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.451747] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f947f439-438a-4e12-8379-349a5e3f0884 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.459965] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea34665c-fd75-41cd-8489-346d5cca21e6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.493419] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf453cda-fab1-44d1-b671-1f2fc32afeb0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.500713] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90352aac-5b60-45b0-8fde-2eb192af7a96 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.514582] env[61985]: DEBUG nova.compute.provider_tree [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.556972] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Successfully created port: 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.595667] env[61985]: DEBUG nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 753.018205] env[61985]: DEBUG nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.346353] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "6316e5ad-6d57-4d3c-baf2-994bb4a3c70d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.818s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.346988] env[61985]: Traceback (most recent call last): [ 753.347063] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 753.347063] env[61985]: self.driver.spawn(context, instance, image_meta, [ 753.347063] env[61985]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 753.347063] env[61985]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.347063] env[61985]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.347063] env[61985]: vm_ref = self.build_virtual_machine(instance, [ 753.347063] env[61985]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.347063] env[61985]: vif_infos = vmwarevif.get_vif_info(self._session, [ 753.347063] env[61985]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.347063] env[61985]: for vif in network_info: [ 753.347063] env[61985]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.347063] env[61985]: return self._sync_wrapper(fn, *args, **kwargs) [ 753.347063] env[61985]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.347063] env[61985]: self.wait() [ 753.347063] env[61985]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.347063] env[61985]: self[:] = self._gt.wait() [ 753.347063] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.347063] env[61985]: return self._exit_event.wait() [ 753.347063] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 753.347063] env[61985]: result = hub.switch() [ 753.347063] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 753.347063] env[61985]: return self.greenlet.switch() [ 753.347063] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.347063] env[61985]: result = function(*args, **kwargs) [ 753.347736] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.347736] env[61985]: return func(*args, **kwargs) [ 753.347736] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.347736] env[61985]: raise e [ 753.347736] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.347736] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 753.347736] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.347736] env[61985]: created_port_ids = self._update_ports_for_instance( [ 753.347736] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.347736] env[61985]: with excutils.save_and_reraise_exception(): [ 753.347736] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.347736] env[61985]: self.force_reraise() [ 753.347736] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.347736] env[61985]: raise self.value [ 753.347736] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.347736] env[61985]: updated_port = self._update_port( [ 753.347736] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.347736] env[61985]: _ensure_no_port_binding_failure(port) [ 753.347736] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.347736] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.347736] env[61985]: nova.exception.PortBindingFailed: Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. [ 753.347736] env[61985]: During handling of the above exception, another exception occurred: [ 753.347736] env[61985]: Traceback (most recent call last): [ 753.347736] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 753.347736] env[61985]: self._build_and_run_instance(context, instance, image, [ 753.347736] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2754, in _build_and_run_instance [ 753.347736] env[61985]: raise exception.RescheduledException( [ 753.347736] env[61985]: nova.exception.RescheduledException: Build of instance 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d was re-scheduled: Binding failed for port e51de715-fd1a-49a0-9cc4-4911cc14b0ec, please check neutron logs for more information. [ 753.348486] env[61985]: During handling of the above exception, another exception occurred: [ 753.348486] env[61985]: Traceback (most recent call last): [ 753.348486] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 753.348486] env[61985]: func(*args, **kwargs) [ 753.348486] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.348486] env[61985]: return func(*args, **kwargs) [ 753.348486] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 753.348486] env[61985]: return f(*args, **kwargs) [ 753.348486] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 753.348486] env[61985]: result = self._do_build_and_run_instance(*args, **kwargs) [ 753.348486] env[61985]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 753.348486] env[61985]: with excutils.save_and_reraise_exception(): [ 753.348486] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.348486] env[61985]: self.force_reraise() [ 753.348486] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.348486] env[61985]: raise self.value [ 753.348486] env[61985]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 753.348486] env[61985]: return f(self, context, *args, **kw) [ 753.348486] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 753.348486] env[61985]: with excutils.save_and_reraise_exception(): [ 753.348486] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.348486] env[61985]: self.force_reraise() [ 753.348486] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.348486] env[61985]: raise self.value [ 753.348486] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 753.348486] env[61985]: return function(self, context, *args, **kwargs) [ 753.348486] env[61985]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 753.348486] env[61985]: return function(self, context, *args, **kwargs) [ 753.348486] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 753.348486] env[61985]: return function(self, context, *args, **kwargs) [ 753.348486] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 753.348486] env[61985]: instance.save() [ 753.349265] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 753.349265] env[61985]: updates, result = self.indirection_api.object_action( [ 753.349265] env[61985]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 753.349265] env[61985]: return cctxt.call(context, 'object_action', objinst=objinst, [ 753.349265] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 753.349265] env[61985]: result = self.transport._send( [ 753.349265] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 753.349265] env[61985]: return self._driver.send(target, ctxt, message, [ 753.349265] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 753.349265] env[61985]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 753.349265] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 753.349265] env[61985]: raise result [ 753.349265] env[61985]: nova.exception_Remote.InstanceNotFound_Remote: Instance 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d could not be found. [ 753.349265] env[61985]: Traceback (most recent call last): [ 753.349265] env[61985]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 753.349265] env[61985]: return getattr(target, method)(*args, **kwargs) [ 753.349265] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 753.349265] env[61985]: return fn(self, *args, **kwargs) [ 753.349265] env[61985]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 753.349265] env[61985]: old_ref, inst_ref = db.instance_update_and_get_original( [ 753.349265] env[61985]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 753.349265] env[61985]: return f(*args, **kwargs) [ 753.349265] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 753.349265] env[61985]: with excutils.save_and_reraise_exception() as ectxt: [ 753.349265] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.349265] env[61985]: self.force_reraise() [ 753.349265] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.349265] env[61985]: raise self.value [ 753.349935] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 753.349935] env[61985]: return f(*args, **kwargs) [ 753.349935] env[61985]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 753.349935] env[61985]: return f(context, *args, **kwargs) [ 753.349935] env[61985]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 753.349935] env[61985]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 753.349935] env[61985]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 753.349935] env[61985]: raise exception.InstanceNotFound(instance_id=uuid) [ 753.349935] env[61985]: nova.exception.InstanceNotFound: Instance 6316e5ad-6d57-4d3c-baf2-994bb4a3c70d could not be found. [ 753.523248] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.940s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.523884] env[61985]: ERROR nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6442d848-93a4-4438-a8bd-534aca1b64f5, please check neutron logs for more information. [ 753.523884] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Traceback (most recent call last): [ 753.523884] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 753.523884] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] self.driver.spawn(context, instance, image_meta, [ 753.523884] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 753.523884] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.523884] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.523884] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] vm_ref = self.build_virtual_machine(instance, [ 753.523884] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.523884] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.523884] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] for vif in network_info: [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] return self._sync_wrapper(fn, *args, **kwargs) [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] self.wait() [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] self[:] = self._gt.wait() [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] return self._exit_event.wait() [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] result = hub.switch() [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 753.524231] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] return self.greenlet.switch() [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] result = function(*args, **kwargs) [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] return func(*args, **kwargs) [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] raise e [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] nwinfo = self.network_api.allocate_for_instance( [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] created_port_ids = self._update_ports_for_instance( [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] with excutils.save_and_reraise_exception(): [ 753.524535] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] self.force_reraise() [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] raise self.value [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] updated_port = self._update_port( [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] _ensure_no_port_binding_failure(port) [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] raise exception.PortBindingFailed(port_id=port['id']) [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] nova.exception.PortBindingFailed: Binding failed for port 6442d848-93a4-4438-a8bd-534aca1b64f5, please check neutron logs for more information. [ 753.524846] env[61985]: ERROR nova.compute.manager [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] [ 753.525121] env[61985]: DEBUG nova.compute.utils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Binding failed for port 6442d848-93a4-4438-a8bd-534aca1b64f5, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 753.525781] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.154s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.529909] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Build of instance 983c68ad-52ca-4eba-9570-3a37d8f5c637 was re-scheduled: Binding failed for port 6442d848-93a4-4438-a8bd-534aca1b64f5, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 753.530364] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 753.530637] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.530737] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquired lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.530911] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.607783] env[61985]: DEBUG nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 753.639144] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 753.639144] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 753.639144] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.639405] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 753.639405] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.639405] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 753.639405] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 753.639405] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 753.639684] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 753.639901] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 753.640142] env[61985]: DEBUG nova.virt.hardware [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 753.641157] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25fe14a-dccd-436a-8d1f-8dd750af469a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.649608] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6fb801b-cd6a-42b4-9431-11a5de75e1c8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.742511] env[61985]: DEBUG nova.compute.manager [req-0cafc2c0-da87-41b0-8d9b-7c8b4a466f17 req-7dd88e7e-e61f-43e7-bb0f-6ab94588301c service nova] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Received event network-changed-5ee9c1b8-1c1c-4ace-a08d-335bbd721c83 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 753.742716] env[61985]: DEBUG nova.compute.manager [req-0cafc2c0-da87-41b0-8d9b-7c8b4a466f17 req-7dd88e7e-e61f-43e7-bb0f-6ab94588301c service nova] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Refreshing instance network info cache due to event network-changed-5ee9c1b8-1c1c-4ace-a08d-335bbd721c83. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 753.742929] env[61985]: DEBUG oslo_concurrency.lockutils [req-0cafc2c0-da87-41b0-8d9b-7c8b4a466f17 req-7dd88e7e-e61f-43e7-bb0f-6ab94588301c service nova] Acquiring lock "refresh_cache-8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.743070] env[61985]: DEBUG oslo_concurrency.lockutils [req-0cafc2c0-da87-41b0-8d9b-7c8b4a466f17 req-7dd88e7e-e61f-43e7-bb0f-6ab94588301c service nova] Acquired lock "refresh_cache-8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.743258] env[61985]: DEBUG nova.network.neutron [req-0cafc2c0-da87-41b0-8d9b-7c8b4a466f17 req-7dd88e7e-e61f-43e7-bb0f-6ab94588301c service nova] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Refreshing network info cache for port 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 753.852999] env[61985]: DEBUG nova.compute.manager [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.915572] env[61985]: ERROR nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83, please check neutron logs for more information. [ 753.915572] env[61985]: ERROR nova.compute.manager Traceback (most recent call last): [ 753.915572] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.915572] env[61985]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 753.915572] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.915572] env[61985]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 753.915572] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.915572] env[61985]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 753.915572] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.915572] env[61985]: ERROR nova.compute.manager self.force_reraise() [ 753.915572] env[61985]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.915572] env[61985]: ERROR nova.compute.manager raise self.value [ 753.915572] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.915572] env[61985]: ERROR nova.compute.manager updated_port = self._update_port( [ 753.915572] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.915572] env[61985]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 753.916032] env[61985]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.916032] env[61985]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 753.916032] env[61985]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83, please check neutron logs for more information. [ 753.916032] env[61985]: ERROR nova.compute.manager [ 753.916032] env[61985]: Traceback (most recent call last): [ 753.916032] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 753.916032] env[61985]: listener.cb(fileno) [ 753.916032] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.916032] env[61985]: result = function(*args, **kwargs) [ 753.916032] env[61985]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.916032] env[61985]: return func(*args, **kwargs) [ 753.916032] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.916032] env[61985]: raise e [ 753.916032] env[61985]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.916032] env[61985]: nwinfo = self.network_api.allocate_for_instance( [ 753.916032] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.916032] env[61985]: created_port_ids = self._update_ports_for_instance( [ 753.916032] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.916032] env[61985]: with excutils.save_and_reraise_exception(): [ 753.916032] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.916032] env[61985]: self.force_reraise() [ 753.916032] env[61985]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.916032] env[61985]: raise self.value [ 753.916032] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.916032] env[61985]: updated_port = self._update_port( [ 753.916032] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.916032] env[61985]: _ensure_no_port_binding_failure(port) [ 753.916032] env[61985]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.916032] env[61985]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.916717] env[61985]: nova.exception.PortBindingFailed: Binding failed for port 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83, please check neutron logs for more information. [ 753.916717] env[61985]: Removing descriptor: 17 [ 753.916717] env[61985]: ERROR nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83, please check neutron logs for more information. [ 753.916717] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Traceback (most recent call last): [ 753.916717] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 753.916717] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] yield resources [ 753.916717] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 753.916717] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] self.driver.spawn(context, instance, image_meta, [ 753.916717] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 753.916717] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.916717] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.916717] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] vm_ref = self.build_virtual_machine(instance, [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] for vif in network_info: [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] return self._sync_wrapper(fn, *args, **kwargs) [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] self.wait() [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] self[:] = self._gt.wait() [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] return self._exit_event.wait() [ 753.917019] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] result = hub.switch() [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] return self.greenlet.switch() [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] result = function(*args, **kwargs) [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] return func(*args, **kwargs) [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] raise e [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] nwinfo = self.network_api.allocate_for_instance( [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.917327] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] created_port_ids = self._update_ports_for_instance( [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] with excutils.save_and_reraise_exception(): [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] self.force_reraise() [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] raise self.value [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] updated_port = self._update_port( [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] _ensure_no_port_binding_failure(port) [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.917719] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] raise exception.PortBindingFailed(port_id=port['id']) [ 753.918022] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] nova.exception.PortBindingFailed: Binding failed for port 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83, please check neutron logs for more information. [ 753.918022] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] [ 753.918022] env[61985]: INFO nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Terminating instance [ 753.919073] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Acquiring lock "refresh_cache-8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.047917] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.167548] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.262394] env[61985]: DEBUG nova.network.neutron [req-0cafc2c0-da87-41b0-8d9b-7c8b4a466f17 req-7dd88e7e-e61f-43e7-bb0f-6ab94588301c service nova] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.338637] env[61985]: DEBUG nova.network.neutron [req-0cafc2c0-da87-41b0-8d9b-7c8b4a466f17 req-7dd88e7e-e61f-43e7-bb0f-6ab94588301c service nova] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.353533] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e79d42-28bd-4fd0-817a-19bc433a2c04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.365989] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f41d311-6754-44da-9461-6045df8667c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.398807] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.398988] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86b5724-73e7-4355-800e-1a9b359a3255 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.406486] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95fc02c-3d27-49d9-b4a4-b014ad41aae1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.419837] env[61985]: DEBUG nova.compute.provider_tree [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.670522] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Releasing lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.670801] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 754.670962] env[61985]: DEBUG nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.671147] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 754.685443] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.843224] env[61985]: DEBUG oslo_concurrency.lockutils [req-0cafc2c0-da87-41b0-8d9b-7c8b4a466f17 req-7dd88e7e-e61f-43e7-bb0f-6ab94588301c service nova] Releasing lock "refresh_cache-8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.843733] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Acquired lock "refresh_cache-8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.843992] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 754.924781] env[61985]: DEBUG nova.scheduler.client.report [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.191159] env[61985]: DEBUG nova.network.neutron [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.365666] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.432624] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.433329] env[61985]: ERROR nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb, please check neutron logs for more information. [ 755.433329] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Traceback (most recent call last): [ 755.433329] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 755.433329] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] self.driver.spawn(context, instance, image_meta, [ 755.433329] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 755.433329] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.433329] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.433329] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] vm_ref = self.build_virtual_machine(instance, [ 755.433329] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.433329] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.433329] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] for vif in network_info: [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] return self._sync_wrapper(fn, *args, **kwargs) [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] self.wait() [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] self[:] = self._gt.wait() [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] return self._exit_event.wait() [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] result = hub.switch() [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 755.433696] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] return self.greenlet.switch() [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] result = function(*args, **kwargs) [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] return func(*args, **kwargs) [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] raise e [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] nwinfo = self.network_api.allocate_for_instance( [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] created_port_ids = self._update_ports_for_instance( [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] with excutils.save_and_reraise_exception(): [ 755.434071] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] self.force_reraise() [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] raise self.value [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] updated_port = self._update_port( [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] _ensure_no_port_binding_failure(port) [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] raise exception.PortBindingFailed(port_id=port['id']) [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] nova.exception.PortBindingFailed: Binding failed for port 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb, please check neutron logs for more information. [ 755.434420] env[61985]: ERROR nova.compute.manager [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] [ 755.434752] env[61985]: DEBUG nova.compute.utils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Binding failed for port 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 755.436958] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.723s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.439957] env[61985]: DEBUG nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Build of instance c8046008-3645-4845-be1e-23f5fefd5dc6 was re-scheduled: Binding failed for port 7a4cbcc9-6d36-409d-9f03-9b4a86a456bb, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 755.440394] env[61985]: DEBUG nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 755.440615] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "refresh_cache-c8046008-3645-4845-be1e-23f5fefd5dc6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.440762] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquired lock "refresh_cache-c8046008-3645-4845-be1e-23f5fefd5dc6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.440920] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 755.529561] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.694862] env[61985]: INFO nova.compute.manager [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Took 1.02 seconds to deallocate network for instance. [ 755.783033] env[61985]: DEBUG nova.compute.manager [req-4e4fb33f-3793-4cb0-8251-d5e13cf5dc07 req-a1a487cf-c35b-47af-8412-77a072e69262 service nova] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Received event network-vif-deleted-5ee9c1b8-1c1c-4ace-a08d-335bbd721c83 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 755.959684] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.032506] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Releasing lock "refresh_cache-8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.032932] env[61985]: DEBUG nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 756.033135] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 756.035413] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66c25c4a-bff8-44df-b1eb-70ae88b6d92b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.045247] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743b863e-7fc6-41e1-94f3-1cf5de901d00 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.064633] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.074029] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c could not be found. [ 756.074154] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 756.074333] env[61985]: INFO nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 756.074658] env[61985]: DEBUG oslo.service.loopingcall [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 756.076106] env[61985]: DEBUG nova.compute.manager [-] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.076219] env[61985]: DEBUG nova.network.neutron [-] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 756.093218] env[61985]: DEBUG nova.network.neutron [-] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.379016] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee9612f-3594-4cc8-a825-107448652a36 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.385896] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65d1f7c-060e-4fe2-8846-1b1ff3c96a9a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.416235] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eba0a8f-6b78-4625-aaa4-688a47321a31 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.423379] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5df7e3-a05d-4702-96bd-9e693b6644a0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.436242] env[61985]: DEBUG nova.compute.provider_tree [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.567081] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Releasing lock "refresh_cache-c8046008-3645-4845-be1e-23f5fefd5dc6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.572696] env[61985]: DEBUG nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 756.572696] env[61985]: DEBUG nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.572696] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 756.596805] env[61985]: DEBUG nova.network.neutron [-] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.598635] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.735989] env[61985]: INFO nova.scheduler.client.report [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Deleted allocations for instance 983c68ad-52ca-4eba-9570-3a37d8f5c637 [ 756.939833] env[61985]: DEBUG nova.scheduler.client.report [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.098266] env[61985]: INFO nova.compute.manager [-] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Took 1.02 seconds to deallocate network for instance. [ 757.100749] env[61985]: DEBUG nova.compute.claims [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Aborting claim: {{(pid=61985) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 757.100923] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.101259] env[61985]: DEBUG nova.network.neutron [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.244423] env[61985]: DEBUG oslo_concurrency.lockutils [None req-92936935-e2c1-40d9-9641-13899a5362a7 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "983c68ad-52ca-4eba-9570-3a37d8f5c637" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.665s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.245698] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "983c68ad-52ca-4eba-9570-3a37d8f5c637" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 4.971s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.245854] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "983c68ad-52ca-4eba-9570-3a37d8f5c637-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.246069] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "983c68ad-52ca-4eba-9570-3a37d8f5c637-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.246287] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "983c68ad-52ca-4eba-9570-3a37d8f5c637-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.248125] env[61985]: INFO nova.compute.manager [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Terminating instance [ 757.252796] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquiring lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.252952] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Acquired lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.253133] env[61985]: DEBUG nova.network.neutron [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 757.445566] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.008s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.446077] env[61985]: ERROR nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6812cb3e-1080-408e-8df0-02f671e82b2f, please check neutron logs for more information. [ 757.446077] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Traceback (most recent call last): [ 757.446077] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 757.446077] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] self.driver.spawn(context, instance, image_meta, [ 757.446077] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 757.446077] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.446077] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.446077] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] vm_ref = self.build_virtual_machine(instance, [ 757.446077] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.446077] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.446077] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] for vif in network_info: [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] return self._sync_wrapper(fn, *args, **kwargs) [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] self.wait() [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] self[:] = self._gt.wait() [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] return self._exit_event.wait() [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] result = hub.switch() [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 757.446481] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] return self.greenlet.switch() [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] result = function(*args, **kwargs) [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] return func(*args, **kwargs) [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] raise e [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] nwinfo = self.network_api.allocate_for_instance( [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] created_port_ids = self._update_ports_for_instance( [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] with excutils.save_and_reraise_exception(): [ 757.446955] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] self.force_reraise() [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] raise self.value [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] updated_port = self._update_port( [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] _ensure_no_port_binding_failure(port) [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] raise exception.PortBindingFailed(port_id=port['id']) [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] nova.exception.PortBindingFailed: Binding failed for port 6812cb3e-1080-408e-8df0-02f671e82b2f, please check neutron logs for more information. [ 757.447348] env[61985]: ERROR nova.compute.manager [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] [ 757.447702] env[61985]: DEBUG nova.compute.utils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Binding failed for port 6812cb3e-1080-408e-8df0-02f671e82b2f, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 757.448116] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.264s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.451747] env[61985]: DEBUG nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Build of instance 3480f0ec-85f4-4881-ad17-d647cd401fbf was re-scheduled: Binding failed for port 6812cb3e-1080-408e-8df0-02f671e82b2f, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 757.452319] env[61985]: DEBUG nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 757.452593] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Acquiring lock "refresh_cache-3480f0ec-85f4-4881-ad17-d647cd401fbf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.452682] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Acquired lock "refresh_cache-3480f0ec-85f4-4881-ad17-d647cd401fbf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.453120] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 757.604347] env[61985]: INFO nova.compute.manager [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: c8046008-3645-4845-be1e-23f5fefd5dc6] Took 1.03 seconds to deallocate network for instance. [ 757.749284] env[61985]: DEBUG nova.compute.manager [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.778938] env[61985]: DEBUG nova.network.neutron [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.882734] env[61985]: DEBUG nova.network.neutron [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.982779] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.072786] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.281239] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.339049] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d424c5b7-9b15-4269-a904-2a1a1fb7bd2f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.348731] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72424697-310e-402f-b1a0-9099611230df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.384644] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b923612-643b-4d8b-a57d-12bb356b9878 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.387620] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Releasing lock "refresh_cache-983c68ad-52ca-4eba-9570-3a37d8f5c637" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.388040] env[61985]: DEBUG nova.compute.manager [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 758.388233] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 758.388525] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-876e75f8-865e-4949-9ac8-b45578d59b40 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.396408] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc33f310-b682-4b19-921a-48ac06351ad7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.402734] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b91cc0-b1eb-4ed5-8b73-19433954f045 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.422337] env[61985]: DEBUG nova.compute.provider_tree [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.430052] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 983c68ad-52ca-4eba-9570-3a37d8f5c637 could not be found. [ 758.430257] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 758.430435] env[61985]: INFO nova.compute.manager [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Took 0.04 seconds to destroy the instance on the hypervisor. [ 758.430672] env[61985]: DEBUG oslo.service.loopingcall [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.431119] env[61985]: DEBUG nova.compute.manager [-] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.431208] env[61985]: DEBUG nova.network.neutron [-] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 758.462733] env[61985]: DEBUG nova.network.neutron [-] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.576084] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Releasing lock "refresh_cache-3480f0ec-85f4-4881-ad17-d647cd401fbf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.576342] env[61985]: DEBUG nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 758.576533] env[61985]: DEBUG nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.576701] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 758.595498] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.636157] env[61985]: INFO nova.scheduler.client.report [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Deleted allocations for instance c8046008-3645-4845-be1e-23f5fefd5dc6 [ 758.933114] env[61985]: DEBUG nova.scheduler.client.report [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.965298] env[61985]: DEBUG nova.network.neutron [-] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.098408] env[61985]: DEBUG nova.network.neutron [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.145788] env[61985]: DEBUG oslo_concurrency.lockutils [None req-37362121-ca86-486d-ab2c-3685cd898d66 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "c8046008-3645-4845-be1e-23f5fefd5dc6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.051s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.440523] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.441204] env[61985]: ERROR nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2a7da93d-f288-4c84-98e1-fc3a5403cb1c, please check neutron logs for more information. [ 759.441204] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Traceback (most recent call last): [ 759.441204] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 759.441204] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] self.driver.spawn(context, instance, image_meta, [ 759.441204] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 759.441204] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.441204] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.441204] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] vm_ref = self.build_virtual_machine(instance, [ 759.441204] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.441204] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.441204] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] for vif in network_info: [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] return self._sync_wrapper(fn, *args, **kwargs) [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] self.wait() [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] self[:] = self._gt.wait() [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] return self._exit_event.wait() [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] result = hub.switch() [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 759.441642] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] return self.greenlet.switch() [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] result = function(*args, **kwargs) [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] return func(*args, **kwargs) [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] raise e [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] nwinfo = self.network_api.allocate_for_instance( [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] created_port_ids = self._update_ports_for_instance( [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] with excutils.save_and_reraise_exception(): [ 759.442196] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] self.force_reraise() [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] raise self.value [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] updated_port = self._update_port( [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] _ensure_no_port_binding_failure(port) [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] raise exception.PortBindingFailed(port_id=port['id']) [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] nova.exception.PortBindingFailed: Binding failed for port 2a7da93d-f288-4c84-98e1-fc3a5403cb1c, please check neutron logs for more information. [ 759.442511] env[61985]: ERROR nova.compute.manager [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] [ 759.442778] env[61985]: DEBUG nova.compute.utils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Binding failed for port 2a7da93d-f288-4c84-98e1-fc3a5403cb1c, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 759.443180] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.674s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.446019] env[61985]: DEBUG nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Build of instance 4a5bd6dd-9d21-43a2-9108-58aed0637ea9 was re-scheduled: Binding failed for port 2a7da93d-f288-4c84-98e1-fc3a5403cb1c, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 759.446537] env[61985]: DEBUG nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 759.446803] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Acquiring lock "refresh_cache-4a5bd6dd-9d21-43a2-9108-58aed0637ea9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.447072] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Acquired lock "refresh_cache-4a5bd6dd-9d21-43a2-9108-58aed0637ea9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.447179] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 759.470163] env[61985]: INFO nova.compute.manager [-] [instance: 983c68ad-52ca-4eba-9570-3a37d8f5c637] Took 1.04 seconds to deallocate network for instance. [ 759.601455] env[61985]: INFO nova.compute.manager [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] [instance: 3480f0ec-85f4-4881-ad17-d647cd401fbf] Took 1.02 seconds to deallocate network for instance. [ 759.647729] env[61985]: DEBUG nova.compute.manager [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.968145] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.078726] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.167775] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.335158] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6859c4e0-4355-4502-bc8e-29d25e8c06d8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.343057] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11673c58-a8cd-4f39-90cd-76978437a3ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.373023] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ee8d91-25cd-46e9-9138-5237038f3869 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.380544] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f1067f-33d6-4534-a506-b4dbd27ae91c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.393562] env[61985]: DEBUG nova.compute.provider_tree [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.504906] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8474822f-1826-4eaa-845c-cf9fbdd583a0 tempest-ListServersNegativeTestJSON-1000445009 tempest-ListServersNegativeTestJSON-1000445009-project-member] Lock "983c68ad-52ca-4eba-9570-3a37d8f5c637" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.259s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.585692] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Releasing lock "refresh_cache-4a5bd6dd-9d21-43a2-9108-58aed0637ea9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.585887] env[61985]: DEBUG nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 760.586093] env[61985]: DEBUG nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.586270] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 760.607899] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.631188] env[61985]: INFO nova.scheduler.client.report [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Deleted allocations for instance 3480f0ec-85f4-4881-ad17-d647cd401fbf [ 760.862946] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquiring lock "f8924f18-7232-4be6-84bf-1ba05b855ffe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.863050] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lock "f8924f18-7232-4be6-84bf-1ba05b855ffe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.897297] env[61985]: DEBUG nova.scheduler.client.report [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.113720] env[61985]: DEBUG nova.network.neutron [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.145343] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8130d84c-6bbb-48b1-94ab-f93644139761 tempest-ServersTestManualDisk-930505171 tempest-ServersTestManualDisk-930505171-project-member] Lock "3480f0ec-85f4-4881-ad17-d647cd401fbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.936s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.363022] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.363227] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.403018] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.403018] env[61985]: ERROR nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc, please check neutron logs for more information. [ 761.403018] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Traceback (most recent call last): [ 761.403018] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 761.403018] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] self.driver.spawn(context, instance, image_meta, [ 761.403018] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 761.403018] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.403018] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.403018] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] vm_ref = self.build_virtual_machine(instance, [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] for vif in network_info: [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] return self._sync_wrapper(fn, *args, **kwargs) [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] self.wait() [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] self[:] = self._gt.wait() [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] return self._exit_event.wait() [ 761.403339] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] result = hub.switch() [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] return self.greenlet.switch() [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] result = function(*args, **kwargs) [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] return func(*args, **kwargs) [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] raise e [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] nwinfo = self.network_api.allocate_for_instance( [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 761.403667] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] created_port_ids = self._update_ports_for_instance( [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] with excutils.save_and_reraise_exception(): [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] self.force_reraise() [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] raise self.value [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] updated_port = self._update_port( [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] _ensure_no_port_binding_failure(port) [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.403980] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] raise exception.PortBindingFailed(port_id=port['id']) [ 761.404293] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] nova.exception.PortBindingFailed: Binding failed for port 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc, please check neutron logs for more information. [ 761.404293] env[61985]: ERROR nova.compute.manager [instance: 9af10e29-12ea-441d-b82a-be6976032c88] [ 761.404293] env[61985]: DEBUG nova.compute.utils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Binding failed for port 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.404713] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.331s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.408076] env[61985]: DEBUG nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Build of instance 9af10e29-12ea-441d-b82a-be6976032c88 was re-scheduled: Binding failed for port 8a26173e-d5a2-4f1a-bd7d-3f032f0067cc, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 761.408571] env[61985]: DEBUG nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 761.408842] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Acquiring lock "refresh_cache-9af10e29-12ea-441d-b82a-be6976032c88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.409038] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Acquired lock "refresh_cache-9af10e29-12ea-441d-b82a-be6976032c88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.409239] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 761.616485] env[61985]: INFO nova.compute.manager [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] [instance: 4a5bd6dd-9d21-43a2-9108-58aed0637ea9] Took 1.03 seconds to deallocate network for instance. [ 761.647924] env[61985]: DEBUG nova.compute.manager [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 761.942496] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.116972] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.169566] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.291258] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3639443-0110-4cff-b2b6-3ac56d748a42 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.299120] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c65348-72f5-4929-b423-c514e82aa3b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.329230] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a27921-5c3e-4ab5-a8f2-af6ce9d361aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.336373] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd612c8-f587-40f6-bee0-b92c21ed2fd2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.350637] env[61985]: DEBUG nova.compute.provider_tree [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.619198] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Releasing lock "refresh_cache-9af10e29-12ea-441d-b82a-be6976032c88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.619681] env[61985]: DEBUG nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 762.619753] env[61985]: DEBUG nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.619926] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 762.642240] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.660211] env[61985]: INFO nova.scheduler.client.report [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Deleted allocations for instance 4a5bd6dd-9d21-43a2-9108-58aed0637ea9 [ 762.855319] env[61985]: DEBUG nova.scheduler.client.report [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.147345] env[61985]: DEBUG nova.network.neutron [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.173848] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c432234c-4c2c-48c4-98e1-4b5dd66633e8 tempest-AttachInterfacesV270Test-2084348482 tempest-AttachInterfacesV270Test-2084348482-project-member] Lock "4a5bd6dd-9d21-43a2-9108-58aed0637ea9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.266s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.360417] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.361077] env[61985]: ERROR nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 10db0b21-b9cc-4931-b892-72fc107d3856, please check neutron logs for more information. [ 763.361077] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Traceback (most recent call last): [ 763.361077] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 763.361077] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] self.driver.spawn(context, instance, image_meta, [ 763.361077] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 763.361077] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.361077] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.361077] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] vm_ref = self.build_virtual_machine(instance, [ 763.361077] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.361077] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.361077] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] for vif in network_info: [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] return self._sync_wrapper(fn, *args, **kwargs) [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] self.wait() [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] self[:] = self._gt.wait() [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] return self._exit_event.wait() [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] result = hub.switch() [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 763.361483] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] return self.greenlet.switch() [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] result = function(*args, **kwargs) [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] return func(*args, **kwargs) [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] raise e [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] nwinfo = self.network_api.allocate_for_instance( [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] created_port_ids = self._update_ports_for_instance( [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] with excutils.save_and_reraise_exception(): [ 763.361802] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] self.force_reraise() [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] raise self.value [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] updated_port = self._update_port( [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] _ensure_no_port_binding_failure(port) [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] raise exception.PortBindingFailed(port_id=port['id']) [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] nova.exception.PortBindingFailed: Binding failed for port 10db0b21-b9cc-4931-b892-72fc107d3856, please check neutron logs for more information. [ 763.362128] env[61985]: ERROR nova.compute.manager [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] [ 763.362421] env[61985]: DEBUG nova.compute.utils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Binding failed for port 10db0b21-b9cc-4931-b892-72fc107d3856, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 763.363113] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.032s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.363335] env[61985]: DEBUG nova.objects.instance [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lazy-loading 'resources' on Instance uuid 213a3e19-5589-4261-96b0-69acfb6798ef {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 763.365014] env[61985]: DEBUG nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Build of instance a0fd407f-7cf0-4c01-8195-399d8cbf62f9 was re-scheduled: Binding failed for port 10db0b21-b9cc-4931-b892-72fc107d3856, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 763.365447] env[61985]: DEBUG nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 763.365912] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Acquiring lock "refresh_cache-a0fd407f-7cf0-4c01-8195-399d8cbf62f9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.365912] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Acquired lock "refresh_cache-a0fd407f-7cf0-4c01-8195-399d8cbf62f9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.366042] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 763.650842] env[61985]: INFO nova.compute.manager [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] [instance: 9af10e29-12ea-441d-b82a-be6976032c88] Took 1.03 seconds to deallocate network for instance. [ 763.677677] env[61985]: DEBUG nova.compute.manager [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 763.889258] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.071818] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.197238] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.283484] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa158aa-605d-45cd-b036-343a3097934c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.291703] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908b23e9-dedf-4aa9-a4e8-8ff3a26fef45 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.323453] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60adf42-a965-430d-a92b-2c8c998501e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.332121] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a320d2-cbcd-4345-bb73-89b538a6e858 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.344792] env[61985]: DEBUG nova.compute.provider_tree [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.580647] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Releasing lock "refresh_cache-a0fd407f-7cf0-4c01-8195-399d8cbf62f9" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.580647] env[61985]: DEBUG nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 764.580647] env[61985]: DEBUG nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.580647] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 764.675720] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.692608] env[61985]: INFO nova.scheduler.client.report [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Deleted allocations for instance 9af10e29-12ea-441d-b82a-be6976032c88 [ 764.849884] env[61985]: DEBUG nova.scheduler.client.report [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.180109] env[61985]: DEBUG nova.network.neutron [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.207120] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8124bb38-23ac-466b-acbf-63e5f7d1fa24 tempest-ServerMetadataNegativeTestJSON-577162010 tempest-ServerMetadataNegativeTestJSON-577162010-project-member] Lock "9af10e29-12ea-441d-b82a-be6976032c88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.299s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.356586] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.991s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.360231] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.777s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.389241] env[61985]: INFO nova.scheduler.client.report [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Deleted allocations for instance 213a3e19-5589-4261-96b0-69acfb6798ef [ 765.685682] env[61985]: INFO nova.compute.manager [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] [instance: a0fd407f-7cf0-4c01-8195-399d8cbf62f9] Took 1.10 seconds to deallocate network for instance. [ 765.711157] env[61985]: DEBUG nova.compute.manager [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 765.904208] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9d27aa5-c199-412c-9aa4-aabe720cf4d9 tempest-ServerShowV247Test-1048534741 tempest-ServerShowV247Test-1048534741-project-member] Lock "213a3e19-5589-4261-96b0-69acfb6798ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.293s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.250895] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.319095] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798e9a4d-5dd7-47e4-aaf7-d72ac75a5b4c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.327642] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab4f1c0-2ccd-4cca-8e3c-37039c572da1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.368400] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762c6d46-04d5-4f76-bb1b-c69b03c75802 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.376529] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219c77b0-0c09-4039-b04c-6975d476b7c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.392098] env[61985]: DEBUG nova.compute.provider_tree [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.734722] env[61985]: INFO nova.scheduler.client.report [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Deleted allocations for instance a0fd407f-7cf0-4c01-8195-399d8cbf62f9 [ 766.893037] env[61985]: DEBUG nova.scheduler.client.report [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.242443] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5facfee3-8a95-4059-98e5-f5fc888e2723 tempest-InstanceActionsV221TestJSON-1154454167 tempest-InstanceActionsV221TestJSON-1154454167-project-member] Lock "a0fd407f-7cf0-4c01-8195-399d8cbf62f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.510s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.398482] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.040s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.399244] env[61985]: ERROR nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2, please check neutron logs for more information. [ 767.399244] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Traceback (most recent call last): [ 767.399244] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 767.399244] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] self.driver.spawn(context, instance, image_meta, [ 767.399244] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 767.399244] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.399244] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.399244] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] vm_ref = self.build_virtual_machine(instance, [ 767.399244] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.399244] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.399244] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] for vif in network_info: [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] return self._sync_wrapper(fn, *args, **kwargs) [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] self.wait() [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] self[:] = self._gt.wait() [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] return self._exit_event.wait() [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] result = hub.switch() [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 767.399578] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] return self.greenlet.switch() [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] result = function(*args, **kwargs) [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] return func(*args, **kwargs) [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] raise e [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] nwinfo = self.network_api.allocate_for_instance( [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] created_port_ids = self._update_ports_for_instance( [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] with excutils.save_and_reraise_exception(): [ 767.399894] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] self.force_reraise() [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] raise self.value [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] updated_port = self._update_port( [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] _ensure_no_port_binding_failure(port) [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] raise exception.PortBindingFailed(port_id=port['id']) [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] nova.exception.PortBindingFailed: Binding failed for port 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2, please check neutron logs for more information. [ 767.400224] env[61985]: ERROR nova.compute.manager [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] [ 767.400499] env[61985]: DEBUG nova.compute.utils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Binding failed for port 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 767.401658] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.471s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.404271] env[61985]: INFO nova.compute.claims [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.408867] env[61985]: DEBUG nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Build of instance 50315640-7609-4fff-b191-aa29f06cb4f1 was re-scheduled: Binding failed for port 7fc59b8e-c7bb-40b1-96d0-f89d40c6d7f2, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 767.411040] env[61985]: DEBUG nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 767.411638] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Acquiring lock "refresh_cache-50315640-7609-4fff-b191-aa29f06cb4f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.414240] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Acquired lock "refresh_cache-50315640-7609-4fff-b191-aa29f06cb4f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.414426] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 767.749765] env[61985]: DEBUG nova.compute.manager [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 767.940596] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.107265] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.283969] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.611337] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Releasing lock "refresh_cache-50315640-7609-4fff-b191-aa29f06cb4f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.611337] env[61985]: DEBUG nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 768.611337] env[61985]: DEBUG nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.611337] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 768.816841] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-479079fe-e23c-49f9-9653-6db3deab375a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.825718] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0090498e-b22e-49e3-b8dd-ed5ffd3e929c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.830691] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.869147] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b61d8e7-8a8c-464b-8db9-dbbf7b3b2516 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.878272] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25d958f-e9c0-493e-938c-054d1df045bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.893992] env[61985]: DEBUG nova.compute.provider_tree [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.014255] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquiring lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.014255] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.334651] env[61985]: DEBUG nova.network.neutron [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.396980] env[61985]: DEBUG nova.scheduler.client.report [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.844255] env[61985]: INFO nova.compute.manager [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] [instance: 50315640-7609-4fff-b191-aa29f06cb4f1] Took 1.23 seconds to deallocate network for instance. [ 769.902959] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.903487] env[61985]: DEBUG nova.compute.manager [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 769.908066] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.510s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.909445] env[61985]: INFO nova.compute.claims [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.410356] env[61985]: DEBUG nova.compute.utils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.415110] env[61985]: DEBUG nova.compute.manager [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 770.415110] env[61985]: DEBUG nova.network.neutron [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 770.491761] env[61985]: DEBUG nova.policy [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f8df9858fb34d4cb63c2d735eb73319', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '697ed07c609f4e1f86d317675a4749a8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 770.544460] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.544687] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.896519] env[61985]: INFO nova.scheduler.client.report [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Deleted allocations for instance 50315640-7609-4fff-b191-aa29f06cb4f1 [ 770.914264] env[61985]: DEBUG nova.compute.manager [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 771.129750] env[61985]: DEBUG nova.network.neutron [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Successfully created port: 4aa18b7b-89b5-4cad-af09-1bbc8f255029 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 771.353141] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61568b6c-8f25-4dd6-901c-6a43741467e4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.362803] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3889822f-98ab-44be-aa04-26bc5d1d836a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.408066] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bdcf95f-b9ff-4fe6-8d88-cae02db3d4ba tempest-ServerRescueTestJSON-1974948395 tempest-ServerRescueTestJSON-1974948395-project-member] Lock "50315640-7609-4fff-b191-aa29f06cb4f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.552s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.408886] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c92ed7-477f-42f2-bc27-e284aadcf3af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.418318] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0221624-3f3c-44bc-8383-86aa51f6083b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.439068] env[61985]: DEBUG nova.compute.provider_tree [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.913321] env[61985]: DEBUG nova.compute.manager [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 771.931036] env[61985]: DEBUG nova.compute.manager [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 771.942651] env[61985]: DEBUG nova.scheduler.client.report [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.965198] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.965457] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.965610] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.965789] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.965934] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.966155] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.966353] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.966513] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.966679] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.966838] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.967051] env[61985]: DEBUG nova.virt.hardware [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.968154] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01352f01-a3a8-42c1-a35f-1d62c40b4be1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.976812] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f376d6d5-d09b-4f2d-81cd-160c42ff286d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.439355] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.447404] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.447933] env[61985]: DEBUG nova.compute.manager [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 772.456056] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.352s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.956929] env[61985]: DEBUG nova.compute.utils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.964018] env[61985]: DEBUG nova.compute.manager [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 772.964018] env[61985]: DEBUG nova.network.neutron [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 773.130993] env[61985]: DEBUG nova.policy [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e2994215f354ca5818a2a3aeac557dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bea886962434d1495fb140b375abde8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 773.404930] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.405329] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.407096] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea8b9d6-584b-4e1e-9f31-c7d03cc2d9f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.415230] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be97cab7-0925-4541-91f7-bda8b7df482f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.446289] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a25482-12db-4d5c-b042-3ff0290a4ec5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.454184] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c6d5e1b-ee14-495b-b8e3-88ed06d7f96f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.469872] env[61985]: DEBUG nova.compute.manager [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 773.472698] env[61985]: DEBUG nova.compute.provider_tree [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.474671] env[61985]: DEBUG nova.compute.manager [req-456063de-27b3-4379-a65f-1eedb89c1f56 req-debc6934-8109-49a6-b216-f8287baea107 service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Received event network-vif-plugged-4aa18b7b-89b5-4cad-af09-1bbc8f255029 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 773.474992] env[61985]: DEBUG oslo_concurrency.lockutils [req-456063de-27b3-4379-a65f-1eedb89c1f56 req-debc6934-8109-49a6-b216-f8287baea107 service nova] Acquiring lock "2322ab72-9841-41fb-9d60-2812baabe108-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.476255] env[61985]: DEBUG oslo_concurrency.lockutils [req-456063de-27b3-4379-a65f-1eedb89c1f56 req-debc6934-8109-49a6-b216-f8287baea107 service nova] Lock "2322ab72-9841-41fb-9d60-2812baabe108-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.476255] env[61985]: DEBUG oslo_concurrency.lockutils [req-456063de-27b3-4379-a65f-1eedb89c1f56 req-debc6934-8109-49a6-b216-f8287baea107 service nova] Lock "2322ab72-9841-41fb-9d60-2812baabe108-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.476255] env[61985]: DEBUG nova.compute.manager [req-456063de-27b3-4379-a65f-1eedb89c1f56 req-debc6934-8109-49a6-b216-f8287baea107 service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] No waiting events found dispatching network-vif-plugged-4aa18b7b-89b5-4cad-af09-1bbc8f255029 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 773.476255] env[61985]: WARNING nova.compute.manager [req-456063de-27b3-4379-a65f-1eedb89c1f56 req-debc6934-8109-49a6-b216-f8287baea107 service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Received unexpected event network-vif-plugged-4aa18b7b-89b5-4cad-af09-1bbc8f255029 for instance with vm_state building and task_state spawning. [ 773.648122] env[61985]: DEBUG nova.network.neutron [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Successfully updated port: 4aa18b7b-89b5-4cad-af09-1bbc8f255029 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 773.803802] env[61985]: DEBUG nova.network.neutron [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Successfully created port: 4afe5581-970d-4d14-b68d-76a76308741e {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.984017] env[61985]: DEBUG nova.scheduler.client.report [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.156488] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.156615] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.156753] env[61985]: DEBUG nova.network.neutron [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 774.270710] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 774.270710] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 774.270710] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 774.270710] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Rebuilding the list of instances to heal {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9936}} [ 774.490582] env[61985]: DEBUG nova.compute.manager [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 774.493304] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.041s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.493912] env[61985]: ERROR nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83, please check neutron logs for more information. [ 774.493912] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Traceback (most recent call last): [ 774.493912] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 774.493912] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] self.driver.spawn(context, instance, image_meta, [ 774.493912] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 774.493912] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 774.493912] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 774.493912] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] vm_ref = self.build_virtual_machine(instance, [ 774.493912] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 774.493912] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] vif_infos = vmwarevif.get_vif_info(self._session, [ 774.493912] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] for vif in network_info: [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] return self._sync_wrapper(fn, *args, **kwargs) [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] self.wait() [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] self[:] = self._gt.wait() [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] return self._exit_event.wait() [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] result = hub.switch() [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 774.494427] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] return self.greenlet.switch() [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] result = function(*args, **kwargs) [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] return func(*args, **kwargs) [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] raise e [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] nwinfo = self.network_api.allocate_for_instance( [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] created_port_ids = self._update_ports_for_instance( [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] with excutils.save_and_reraise_exception(): [ 774.494927] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] self.force_reraise() [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] raise self.value [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] updated_port = self._update_port( [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] _ensure_no_port_binding_failure(port) [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] raise exception.PortBindingFailed(port_id=port['id']) [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] nova.exception.PortBindingFailed: Binding failed for port 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83, please check neutron logs for more information. [ 774.495421] env[61985]: ERROR nova.compute.manager [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] [ 774.495811] env[61985]: DEBUG nova.compute.utils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Binding failed for port 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83, please check neutron logs for more information. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 774.496315] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.219s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.497728] env[61985]: INFO nova.compute.claims [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.500301] env[61985]: DEBUG nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Build of instance 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c was re-scheduled: Binding failed for port 5ee9c1b8-1c1c-4ace-a08d-335bbd721c83, please check neutron logs for more information. {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 774.500804] env[61985]: DEBUG nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Unplugging VIFs for instance {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 774.500959] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Acquiring lock "refresh_cache-8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.501125] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Acquired lock "refresh_cache-8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.501286] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 774.523707] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.524047] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.524158] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.524352] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.524500] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.524673] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.524892] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.525068] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.525281] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.525665] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.525731] env[61985]: DEBUG nova.virt.hardware [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.526834] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e549eb1c-429c-462a-a6f3-f96d336aa958 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.535530] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe3c12a-2472-4644-b796-3b9424eea99c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.722162] env[61985]: DEBUG nova.network.neutron [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.776081] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 774.776251] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 774.776383] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Didn't find any instances for network info cache update. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10018}} [ 774.776575] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 774.776735] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 774.776871] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 774.777030] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 774.777178] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 774.777319] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 774.777443] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 774.777611] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.020974] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.107129] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.112966] env[61985]: DEBUG nova.network.neutron [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance_info_cache with network_info: [{"id": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "address": "fa:16:3e:12:45:ef", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aa18b7b-89", "ovs_interfaceid": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.285397] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.540632] env[61985]: DEBUG nova.compute.manager [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Received event network-changed-4aa18b7b-89b5-4cad-af09-1bbc8f255029 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 775.541042] env[61985]: DEBUG nova.compute.manager [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Refreshing instance network info cache due to event network-changed-4aa18b7b-89b5-4cad-af09-1bbc8f255029. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 775.541701] env[61985]: DEBUG oslo_concurrency.lockutils [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] Acquiring lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.569990] env[61985]: DEBUG nova.network.neutron [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Successfully updated port: 4afe5581-970d-4d14-b68d-76a76308741e {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 775.610993] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Releasing lock "refresh_cache-8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.611230] env[61985]: DEBUG nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61985) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 775.611416] env[61985]: DEBUG nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.611579] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 775.615506] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.618124] env[61985]: DEBUG nova.compute.manager [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Instance network_info: |[{"id": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "address": "fa:16:3e:12:45:ef", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aa18b7b-89", "ovs_interfaceid": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 775.618124] env[61985]: DEBUG oslo_concurrency.lockutils [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] Acquired lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.618402] env[61985]: DEBUG nova.network.neutron [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Refreshing network info cache for port 4aa18b7b-89b5-4cad-af09-1bbc8f255029 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 775.618402] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:45:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4aa18b7b-89b5-4cad-af09-1bbc8f255029', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 775.631051] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating folder: Project (697ed07c609f4e1f86d317675a4749a8). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 775.634119] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.639117] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53c6cfa2-4345-4da5-83ad-f31caf826d23 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.651530] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Created folder: Project (697ed07c609f4e1f86d317675a4749a8) in parent group-v211285. [ 775.651753] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating folder: Instances. Parent ref: group-v211298. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 775.655960] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a44d2a5f-5726-433b-8c28-f370e363effd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.668173] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Created folder: Instances in parent group-v211298. [ 775.668173] env[61985]: DEBUG oslo.service.loopingcall [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.669451] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 775.669451] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6efad76-1c88-460a-9313-224ca50787df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.695610] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 775.695610] env[61985]: value = "task-935659" [ 775.695610] env[61985]: _type = "Task" [ 775.695610] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.712311] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935659, 'name': CreateVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.942148] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e238cd54-5503-4aa4-87ea-85f1562e7312 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.950184] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a073021-ab3f-4683-9139-4eca8c0e6d6e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.984537] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4cc8d9-6e08-4ac3-a45d-d45fa1567c93 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.993044] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc152ab-010a-4bdf-97a6-0443d6f1cb80 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.006701] env[61985]: DEBUG nova.compute.provider_tree [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.035703] env[61985]: DEBUG nova.network.neutron [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updated VIF entry in instance network info cache for port 4aa18b7b-89b5-4cad-af09-1bbc8f255029. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 776.036170] env[61985]: DEBUG nova.network.neutron [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance_info_cache with network_info: [{"id": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "address": "fa:16:3e:12:45:ef", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aa18b7b-89", "ovs_interfaceid": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.075798] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "refresh_cache-4add2608-8187-42a2-9897-3e5a9049757e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.075976] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquired lock "refresh_cache-4add2608-8187-42a2-9897-3e5a9049757e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.076129] env[61985]: DEBUG nova.network.neutron [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.143307] env[61985]: DEBUG nova.network.neutron [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.205138] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935659, 'name': CreateVM_Task, 'duration_secs': 0.350536} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.205326] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 776.212518] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.212693] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.213022] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 776.213269] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c863d98-bbc5-4fe7-ab4c-ea2abeeb7a95 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.218334] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 776.218334] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52eee5b1-e954-6c22-1fab-cbaaab16375f" [ 776.218334] env[61985]: _type = "Task" [ 776.218334] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.226607] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52eee5b1-e954-6c22-1fab-cbaaab16375f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.513023] env[61985]: DEBUG nova.scheduler.client.report [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.539136] env[61985]: DEBUG oslo_concurrency.lockutils [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] Releasing lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.539136] env[61985]: DEBUG nova.compute.manager [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Received event network-vif-plugged-4afe5581-970d-4d14-b68d-76a76308741e {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 776.539295] env[61985]: DEBUG oslo_concurrency.lockutils [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] Acquiring lock "4add2608-8187-42a2-9897-3e5a9049757e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.539932] env[61985]: DEBUG oslo_concurrency.lockutils [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] Lock "4add2608-8187-42a2-9897-3e5a9049757e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.539932] env[61985]: DEBUG oslo_concurrency.lockutils [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] Lock "4add2608-8187-42a2-9897-3e5a9049757e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.539932] env[61985]: DEBUG nova.compute.manager [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] No waiting events found dispatching network-vif-plugged-4afe5581-970d-4d14-b68d-76a76308741e {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 776.540128] env[61985]: WARNING nova.compute.manager [req-99369da0-ef7f-4c46-9b8e-e06b4ef341db req-2257c729-a8e6-4541-bd3f-d2b0ce4a736b service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Received unexpected event network-vif-plugged-4afe5581-970d-4d14-b68d-76a76308741e for instance with vm_state building and task_state spawning. [ 776.607204] env[61985]: DEBUG nova.network.neutron [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.645730] env[61985]: INFO nova.compute.manager [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] [instance: 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c] Took 1.03 seconds to deallocate network for instance. [ 776.728904] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52eee5b1-e954-6c22-1fab-cbaaab16375f, 'name': SearchDatastore_Task, 'duration_secs': 0.011582} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.729254] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.729489] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 776.729723] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.729890] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.730092] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 776.730348] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed6027ad-13a2-4881-8b36-440d223fe97b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.737561] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 776.737749] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 776.738466] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f04cf068-2ff8-42d4-b076-d071c28e7c8d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.743878] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 776.743878] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527b58d3-80a8-2b50-15e6-1d95ed178438" [ 776.743878] env[61985]: _type = "Task" [ 776.743878] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.750990] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527b58d3-80a8-2b50-15e6-1d95ed178438, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.760869] env[61985]: DEBUG nova.network.neutron [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Updating instance_info_cache with network_info: [{"id": "4afe5581-970d-4d14-b68d-76a76308741e", "address": "fa:16:3e:96:80:59", "network": {"id": "09dee7f0-59e0-4b8d-bf1d-b1a6c998e2d6", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-366451205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bea886962434d1495fb140b375abde8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4afe5581-97", "ovs_interfaceid": "4afe5581-970d-4d14-b68d-76a76308741e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.018392] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.018700] env[61985]: DEBUG nova.compute.manager [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 777.021326] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.854s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.022840] env[61985]: INFO nova.compute.claims [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.253940] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527b58d3-80a8-2b50-15e6-1d95ed178438, 'name': SearchDatastore_Task, 'duration_secs': 0.008228} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.254726] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67efc64c-f4a5-49a0-8e7d-610ff9a6ea5e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.260336] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 777.260336] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e50b28-5ea0-c601-a354-7164709070b3" [ 777.260336] env[61985]: _type = "Task" [ 777.260336] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.263467] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Releasing lock "refresh_cache-4add2608-8187-42a2-9897-3e5a9049757e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.263757] env[61985]: DEBUG nova.compute.manager [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Instance network_info: |[{"id": "4afe5581-970d-4d14-b68d-76a76308741e", "address": "fa:16:3e:96:80:59", "network": {"id": "09dee7f0-59e0-4b8d-bf1d-b1a6c998e2d6", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-366451205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bea886962434d1495fb140b375abde8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4afe5581-97", "ovs_interfaceid": "4afe5581-970d-4d14-b68d-76a76308741e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 777.264136] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:80:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4afe5581-970d-4d14-b68d-76a76308741e', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 777.271698] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Creating folder: Project (8bea886962434d1495fb140b375abde8). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 777.272303] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9bf97c24-5641-47ba-abe4-44a6f952a41e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.277866] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e50b28-5ea0-c601-a354-7164709070b3, 'name': SearchDatastore_Task, 'duration_secs': 0.008251} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.278197] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.278440] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 2322ab72-9841-41fb-9d60-2812baabe108/2322ab72-9841-41fb-9d60-2812baabe108.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 777.278673] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e6d8f8f-dbed-44cb-a285-286a347d36b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.284877] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 777.284877] env[61985]: value = "task-935661" [ 777.284877] env[61985]: _type = "Task" [ 777.284877] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.289078] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Created folder: Project (8bea886962434d1495fb140b375abde8) in parent group-v211285. [ 777.289345] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Creating folder: Instances. Parent ref: group-v211301. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 777.289882] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34abb66c-7863-42ea-9482-9d992d84b7f0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.294596] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935661, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.298099] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Created folder: Instances in parent group-v211301. [ 777.298350] env[61985]: DEBUG oslo.service.loopingcall [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.298579] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 777.298793] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a838e51a-1e6c-4f87-8cf3-ead30ef22d33 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.318260] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 777.318260] env[61985]: value = "task-935663" [ 777.318260] env[61985]: _type = "Task" [ 777.318260] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.325359] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935663, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.527134] env[61985]: DEBUG nova.compute.utils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.531062] env[61985]: DEBUG nova.compute.manager [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 777.531062] env[61985]: DEBUG nova.network.neutron [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 777.663059] env[61985]: DEBUG nova.policy [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '248945e574d34c10a4f2387688673613', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '643e8788ab5746cb9d3259031c275261', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 777.679152] env[61985]: DEBUG nova.compute.manager [req-e6c42aa7-9f03-4fab-95f0-e97a59d0849c req-41418e71-56f7-4bd4-917a-e3c64ee3cca3 service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Received event network-changed-4afe5581-970d-4d14-b68d-76a76308741e {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 777.679361] env[61985]: DEBUG nova.compute.manager [req-e6c42aa7-9f03-4fab-95f0-e97a59d0849c req-41418e71-56f7-4bd4-917a-e3c64ee3cca3 service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Refreshing instance network info cache due to event network-changed-4afe5581-970d-4d14-b68d-76a76308741e. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 777.679574] env[61985]: DEBUG oslo_concurrency.lockutils [req-e6c42aa7-9f03-4fab-95f0-e97a59d0849c req-41418e71-56f7-4bd4-917a-e3c64ee3cca3 service nova] Acquiring lock "refresh_cache-4add2608-8187-42a2-9897-3e5a9049757e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.679719] env[61985]: DEBUG oslo_concurrency.lockutils [req-e6c42aa7-9f03-4fab-95f0-e97a59d0849c req-41418e71-56f7-4bd4-917a-e3c64ee3cca3 service nova] Acquired lock "refresh_cache-4add2608-8187-42a2-9897-3e5a9049757e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.679893] env[61985]: DEBUG nova.network.neutron [req-e6c42aa7-9f03-4fab-95f0-e97a59d0849c req-41418e71-56f7-4bd4-917a-e3c64ee3cca3 service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Refreshing network info cache for port 4afe5581-970d-4d14-b68d-76a76308741e {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 777.695692] env[61985]: INFO nova.scheduler.client.report [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Deleted allocations for instance 8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c [ 777.797763] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935661, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470578} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.797763] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 2322ab72-9841-41fb-9d60-2812baabe108/2322ab72-9841-41fb-9d60-2812baabe108.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 777.797763] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 777.797763] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-851aac2c-7d3a-4b94-8b5d-72f3835439c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.803870] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 777.803870] env[61985]: value = "task-935664" [ 777.803870] env[61985]: _type = "Task" [ 777.803870] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.810622] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935664, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.827013] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935663, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.034430] env[61985]: DEBUG nova.compute.manager [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 778.150387] env[61985]: DEBUG nova.network.neutron [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Successfully created port: adde3d48-ebd4-450d-920a-fb93eeff4a8b {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.204847] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ffef45a0-bb17-4e3a-9755-5ee49d53afcc tempest-ServersTestFqdnHostnames-1626308606 tempest-ServersTestFqdnHostnames-1626308606-project-member] Lock "8efc67a0-68e8-4d3e-9cf2-d0eb1a49611c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.668s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.315554] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935664, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.166935} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.317353] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 778.318385] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e39d1d-b31b-4cc6-b501-179f96a5ac2f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.328544] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935663, 'name': CreateVM_Task, 'duration_secs': 0.690349} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.337321] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 778.347189] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 2322ab72-9841-41fb-9d60-2812baabe108/2322ab72-9841-41fb-9d60-2812baabe108.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 778.351585] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.351751] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.352071] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 778.352310] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-637ed545-4b89-4bae-83aa-a26fa9a10253 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.366296] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44d79527-1bc1-4399-b684-bb9035a0343e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.372773] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for the task: (returnval){ [ 778.372773] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5271d3b7-2d7b-14df-6250-179688b6e3af" [ 778.372773] env[61985]: _type = "Task" [ 778.372773] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.374131] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 778.374131] env[61985]: value = "task-935665" [ 778.374131] env[61985]: _type = "Task" [ 778.374131] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.387961] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5271d3b7-2d7b-14df-6250-179688b6e3af, 'name': SearchDatastore_Task, 'duration_secs': 0.009598} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.391055] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.391295] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 778.391516] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.391652] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.391823] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 778.392088] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935665, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.392367] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5150e543-7336-4013-aef8-57291c41c2c8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.405022] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 778.405022] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 778.405022] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-961f9017-c4a3-417a-bd9d-a93e31c74f17 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.413755] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for the task: (returnval){ [ 778.413755] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5218480a-7acd-3177-9d47-e01166ce95ba" [ 778.413755] env[61985]: _type = "Task" [ 778.413755] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.422383] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5218480a-7acd-3177-9d47-e01166ce95ba, 'name': SearchDatastore_Task, 'duration_secs': 0.008555} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.423471] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-031380c1-abbd-4c8c-8e8f-9e450e6d6bb9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.430960] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for the task: (returnval){ [ 778.430960] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5269e678-c9a1-9c38-6ced-ea848fcdd1f3" [ 778.430960] env[61985]: _type = "Task" [ 778.430960] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.438550] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5269e678-c9a1-9c38-6ced-ea848fcdd1f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.440811] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4577595-9036-45d8-b4ca-88f37794e02c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.446851] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af21e9b-a979-4d61-960f-6cf3c9e4f052 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.480146] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c046bb-2ad5-43d3-a0a9-1247fa52dc74 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.487736] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213df9e0-64cd-48e0-8b8f-96a73a58dc8f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.501397] env[61985]: DEBUG nova.compute.provider_tree [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.711973] env[61985]: DEBUG nova.compute.manager [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 778.735251] env[61985]: DEBUG nova.network.neutron [req-e6c42aa7-9f03-4fab-95f0-e97a59d0849c req-41418e71-56f7-4bd4-917a-e3c64ee3cca3 service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Updated VIF entry in instance network info cache for port 4afe5581-970d-4d14-b68d-76a76308741e. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 778.735595] env[61985]: DEBUG nova.network.neutron [req-e6c42aa7-9f03-4fab-95f0-e97a59d0849c req-41418e71-56f7-4bd4-917a-e3c64ee3cca3 service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Updating instance_info_cache with network_info: [{"id": "4afe5581-970d-4d14-b68d-76a76308741e", "address": "fa:16:3e:96:80:59", "network": {"id": "09dee7f0-59e0-4b8d-bf1d-b1a6c998e2d6", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-366451205-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bea886962434d1495fb140b375abde8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4afe5581-97", "ovs_interfaceid": "4afe5581-970d-4d14-b68d-76a76308741e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.890352] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935665, 'name': ReconfigVM_Task, 'duration_secs': 0.273174} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.890627] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 2322ab72-9841-41fb-9d60-2812baabe108/2322ab72-9841-41fb-9d60-2812baabe108.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 778.891275] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70dd0dad-f8ff-4c5a-8cb0-584a740d3ce6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.897971] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 778.897971] env[61985]: value = "task-935666" [ 778.897971] env[61985]: _type = "Task" [ 778.897971] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.905964] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935666, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.942452] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5269e678-c9a1-9c38-6ced-ea848fcdd1f3, 'name': SearchDatastore_Task, 'duration_secs': 0.008183} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.942722] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.942986] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 4add2608-8187-42a2-9897-3e5a9049757e/4add2608-8187-42a2-9897-3e5a9049757e.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 778.943252] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2625dfb-a815-4d9b-a645-827d5f7fbd71 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.949314] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for the task: (returnval){ [ 778.949314] env[61985]: value = "task-935667" [ 778.949314] env[61985]: _type = "Task" [ 778.949314] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.957764] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935667, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.005093] env[61985]: DEBUG nova.scheduler.client.report [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.050880] env[61985]: DEBUG nova.compute.manager [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 779.077633] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.077841] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.078047] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.078187] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.078336] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.078494] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.078794] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.078952] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.079204] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.079452] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.079452] env[61985]: DEBUG nova.virt.hardware [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.080375] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2f9f47-dfe0-4113-a737-64a2f967e618 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.088381] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d292675-0900-43ab-9668-890b8b287de9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.237710] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.238281] env[61985]: DEBUG oslo_concurrency.lockutils [req-e6c42aa7-9f03-4fab-95f0-e97a59d0849c req-41418e71-56f7-4bd4-917a-e3c64ee3cca3 service nova] Releasing lock "refresh_cache-4add2608-8187-42a2-9897-3e5a9049757e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.410193] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935666, 'name': Rename_Task, 'duration_secs': 0.145564} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.410590] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 779.410852] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f364dcc-9aef-42d8-95f5-95e84cf889b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.417666] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 779.417666] env[61985]: value = "task-935668" [ 779.417666] env[61985]: _type = "Task" [ 779.417666] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.426429] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935668, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.459158] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935667, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475781} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.459455] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 4add2608-8187-42a2-9897-3e5a9049757e/4add2608-8187-42a2-9897-3e5a9049757e.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 779.459710] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 779.460016] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b883149-44e2-457e-90df-58fa23ff6c7c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.467722] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for the task: (returnval){ [ 779.467722] env[61985]: value = "task-935669" [ 779.467722] env[61985]: _type = "Task" [ 779.467722] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.476273] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935669, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.510373] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.510967] env[61985]: DEBUG nova.compute.manager [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 779.513941] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.344s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.515419] env[61985]: INFO nova.compute.claims [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.928023] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935668, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.978879] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935669, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065717} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.979163] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 779.979944] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aaac1d0-e553-489c-b133-216e5c37c1fe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.003088] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 4add2608-8187-42a2-9897-3e5a9049757e/4add2608-8187-42a2-9897-3e5a9049757e.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.003392] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec7028e4-8aee-444d-94a9-01273a695242 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.021255] env[61985]: DEBUG nova.compute.utils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.022610] env[61985]: DEBUG nova.compute.manager [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Not allocating networking since 'none' was specified. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 780.031160] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for the task: (returnval){ [ 780.031160] env[61985]: value = "task-935670" [ 780.031160] env[61985]: _type = "Task" [ 780.031160] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.041464] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935670, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.360670] env[61985]: DEBUG nova.network.neutron [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Successfully updated port: adde3d48-ebd4-450d-920a-fb93eeff4a8b {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 780.433273] env[61985]: DEBUG oslo_vmware.api [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935668, 'name': PowerOnVM_Task, 'duration_secs': 0.541175} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.434379] env[61985]: DEBUG nova.compute.manager [req-9982e53a-65fc-46a3-b5af-626166cb602c req-509bb29d-4c74-472a-9a4e-9dec5cadad3c service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Received event network-vif-plugged-adde3d48-ebd4-450d-920a-fb93eeff4a8b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 780.434589] env[61985]: DEBUG oslo_concurrency.lockutils [req-9982e53a-65fc-46a3-b5af-626166cb602c req-509bb29d-4c74-472a-9a4e-9dec5cadad3c service nova] Acquiring lock "fea0cf39-e851-409f-86f5-31cc128a44dc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.434796] env[61985]: DEBUG oslo_concurrency.lockutils [req-9982e53a-65fc-46a3-b5af-626166cb602c req-509bb29d-4c74-472a-9a4e-9dec5cadad3c service nova] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.435094] env[61985]: DEBUG oslo_concurrency.lockutils [req-9982e53a-65fc-46a3-b5af-626166cb602c req-509bb29d-4c74-472a-9a4e-9dec5cadad3c service nova] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.435158] env[61985]: DEBUG nova.compute.manager [req-9982e53a-65fc-46a3-b5af-626166cb602c req-509bb29d-4c74-472a-9a4e-9dec5cadad3c service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] No waiting events found dispatching network-vif-plugged-adde3d48-ebd4-450d-920a-fb93eeff4a8b {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 780.435323] env[61985]: WARNING nova.compute.manager [req-9982e53a-65fc-46a3-b5af-626166cb602c req-509bb29d-4c74-472a-9a4e-9dec5cadad3c service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Received unexpected event network-vif-plugged-adde3d48-ebd4-450d-920a-fb93eeff4a8b for instance with vm_state building and task_state spawning. [ 780.435694] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 780.435904] env[61985]: INFO nova.compute.manager [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Took 8.51 seconds to spawn the instance on the hypervisor. [ 780.436149] env[61985]: DEBUG nova.compute.manager [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 780.437818] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1093a12e-508a-409f-8e18-b6a7e6f0fd83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.528694] env[61985]: DEBUG nova.compute.manager [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 780.541870] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935670, 'name': ReconfigVM_Task, 'duration_secs': 0.270035} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.542222] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 4add2608-8187-42a2-9897-3e5a9049757e/4add2608-8187-42a2-9897-3e5a9049757e.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 780.542930] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2ac61ec-adeb-4941-9839-245e87e53b6c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.551831] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for the task: (returnval){ [ 780.551831] env[61985]: value = "task-935671" [ 780.551831] env[61985]: _type = "Task" [ 780.551831] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.560837] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935671, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.851094] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84f3b24-a7fb-43b1-a63d-9bed10c8c8b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.859768] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44cba256-ddb3-4a1b-b27d-b7307632c8fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.863110] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.863242] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.863388] env[61985]: DEBUG nova.network.neutron [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 780.896594] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed42d570-ad7f-409d-ab78-14a30b9fa589 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.904939] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116b09af-d9bc-4ee8-a5ad-9199f62e547e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.919754] env[61985]: DEBUG nova.compute.provider_tree [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 780.959730] env[61985]: INFO nova.compute.manager [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Took 29.04 seconds to build instance. [ 781.060505] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935671, 'name': Rename_Task, 'duration_secs': 0.136509} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.061239] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 781.061545] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec0b4067-ce8d-4f99-b508-95c1efa917ce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.068386] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for the task: (returnval){ [ 781.068386] env[61985]: value = "task-935672" [ 781.068386] env[61985]: _type = "Task" [ 781.068386] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.076156] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.438976] env[61985]: DEBUG nova.network.neutron [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.442468] env[61985]: ERROR nova.scheduler.client.report [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [req-81938552-d275-41ae-b6a1-67ede38667f8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-81938552-d275-41ae-b6a1-67ede38667f8"}]} [ 781.461339] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f1912505-07b3-4e02-b813-e650b7ea5c44 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.429s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.463587] env[61985]: DEBUG nova.scheduler.client.report [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 781.483168] env[61985]: DEBUG nova.scheduler.client.report [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 781.483442] env[61985]: DEBUG nova.compute.provider_tree [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 781.495736] env[61985]: DEBUG nova.scheduler.client.report [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 781.518529] env[61985]: DEBUG nova.scheduler.client.report [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 781.548734] env[61985]: DEBUG nova.compute.manager [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 781.589588] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935672, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.591809] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.592035] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.592209] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.592421] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.592561] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.592731] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.593333] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.593684] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.594278] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.594467] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.594814] env[61985]: DEBUG nova.virt.hardware [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.595693] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82430cb-e180-47f2-9ab4-96a16cd78688 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.607529] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb67847-ddcc-450d-ac83-654066f98aa8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.626142] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.631749] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Creating folder: Project (2196a105b50c4a129533fc6e1703e7c8). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.638404] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c9d29c0-36e2-4d63-856f-944214b2fe52 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.651249] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Created folder: Project (2196a105b50c4a129533fc6e1703e7c8) in parent group-v211285. [ 781.651249] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Creating folder: Instances. Parent ref: group-v211304. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.651249] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4f8f83c-c53f-4331-9625-6e352f56b6e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.660377] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Created folder: Instances in parent group-v211304. [ 781.663021] env[61985]: DEBUG oslo.service.loopingcall [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.663021] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 781.663021] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1292996-bf78-49b3-8467-11f5ec3a374d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.681739] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.681739] env[61985]: value = "task-935675" [ 781.681739] env[61985]: _type = "Task" [ 781.681739] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.691681] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935675, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.791842] env[61985]: DEBUG nova.network.neutron [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance_info_cache with network_info: [{"id": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "address": "fa:16:3e:7b:a1:a8", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadde3d48-eb", "ovs_interfaceid": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.902788] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53164bec-35b6-4438-ac43-2de27197bb26 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.910347] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177b9ba0-9652-409c-b79c-2ef772b36814 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.941266] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08aff363-257e-45de-94b4-bfa265069d13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.949176] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c93b11-7e40-4e91-8e84-d7af16bea825 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.965503] env[61985]: DEBUG nova.compute.provider_tree [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 781.969523] env[61985]: DEBUG nova.compute.manager [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 782.081500] env[61985]: DEBUG oslo_vmware.api [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935672, 'name': PowerOnVM_Task, 'duration_secs': 0.617376} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.081500] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 782.081500] env[61985]: INFO nova.compute.manager [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Took 7.59 seconds to spawn the instance on the hypervisor. [ 782.081699] env[61985]: DEBUG nova.compute.manager [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.082425] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d269d38-5f09-422e-92dc-702e018d91d8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.193703] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935675, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.299585] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.299937] env[61985]: DEBUG nova.compute.manager [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Instance network_info: |[{"id": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "address": "fa:16:3e:7b:a1:a8", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadde3d48-eb", "ovs_interfaceid": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 782.300387] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:a1:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e1049e8-c06b-4c93-a9e1-2cbb530f3f95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'adde3d48-ebd4-450d-920a-fb93eeff4a8b', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.307993] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Creating folder: Project (643e8788ab5746cb9d3259031c275261). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 782.308262] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fbe1fca0-5e94-4580-b2ce-1f959a2df62b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.320661] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Created folder: Project (643e8788ab5746cb9d3259031c275261) in parent group-v211285. [ 782.320836] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Creating folder: Instances. Parent ref: group-v211307. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 782.321060] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f42974fb-e609-4ca2-a5ec-a150650a3461 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.329527] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Created folder: Instances in parent group-v211307. [ 782.329768] env[61985]: DEBUG oslo.service.loopingcall [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.333017] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 782.333017] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d066c137-e63d-46c4-988e-f3e6493e6906 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.352676] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.352676] env[61985]: value = "task-935678" [ 782.352676] env[61985]: _type = "Task" [ 782.352676] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.362874] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935678, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.464350] env[61985]: DEBUG nova.compute.manager [req-6eb40481-5727-460c-834e-df9733df502e req-f6b6930a-3523-4a03-a70e-5c368d9416cc service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Received event network-changed-adde3d48-ebd4-450d-920a-fb93eeff4a8b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 782.464350] env[61985]: DEBUG nova.compute.manager [req-6eb40481-5727-460c-834e-df9733df502e req-f6b6930a-3523-4a03-a70e-5c368d9416cc service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Refreshing instance network info cache due to event network-changed-adde3d48-ebd4-450d-920a-fb93eeff4a8b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 782.464350] env[61985]: DEBUG oslo_concurrency.lockutils [req-6eb40481-5727-460c-834e-df9733df502e req-f6b6930a-3523-4a03-a70e-5c368d9416cc service nova] Acquiring lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.464656] env[61985]: DEBUG oslo_concurrency.lockutils [req-6eb40481-5727-460c-834e-df9733df502e req-f6b6930a-3523-4a03-a70e-5c368d9416cc service nova] Acquired lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.464656] env[61985]: DEBUG nova.network.neutron [req-6eb40481-5727-460c-834e-df9733df502e req-f6b6930a-3523-4a03-a70e-5c368d9416cc service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Refreshing network info cache for port adde3d48-ebd4-450d-920a-fb93eeff4a8b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.496054] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.501877] env[61985]: DEBUG nova.scheduler.client.report [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 69 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 782.502137] env[61985]: DEBUG nova.compute.provider_tree [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 69 to 70 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 782.502312] env[61985]: DEBUG nova.compute.provider_tree [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 782.532953] env[61985]: DEBUG nova.compute.manager [req-83ef90e5-538e-4d7f-a9bb-90ad4e69abcb req-58f1fe26-7bc4-4966-8253-814301aec89d service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Received event network-changed-4aa18b7b-89b5-4cad-af09-1bbc8f255029 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 782.533409] env[61985]: DEBUG nova.compute.manager [req-83ef90e5-538e-4d7f-a9bb-90ad4e69abcb req-58f1fe26-7bc4-4966-8253-814301aec89d service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Refreshing instance network info cache due to event network-changed-4aa18b7b-89b5-4cad-af09-1bbc8f255029. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 782.533409] env[61985]: DEBUG oslo_concurrency.lockutils [req-83ef90e5-538e-4d7f-a9bb-90ad4e69abcb req-58f1fe26-7bc4-4966-8253-814301aec89d service nova] Acquiring lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.533513] env[61985]: DEBUG oslo_concurrency.lockutils [req-83ef90e5-538e-4d7f-a9bb-90ad4e69abcb req-58f1fe26-7bc4-4966-8253-814301aec89d service nova] Acquired lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.533648] env[61985]: DEBUG nova.network.neutron [req-83ef90e5-538e-4d7f-a9bb-90ad4e69abcb req-58f1fe26-7bc4-4966-8253-814301aec89d service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Refreshing network info cache for port 4aa18b7b-89b5-4cad-af09-1bbc8f255029 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.603173] env[61985]: INFO nova.compute.manager [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Took 28.24 seconds to build instance. [ 782.695292] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935675, 'name': CreateVM_Task, 'duration_secs': 0.618057} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.695604] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 782.696112] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.696572] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.697068] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.697437] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f22a1b38-3a39-4c35-854c-a5d9078a6f89 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.702558] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 782.702558] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52928ae0-7c43-cb5e-44b7-245ec0cf8f74" [ 782.702558] env[61985]: _type = "Task" [ 782.702558] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.710965] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52928ae0-7c43-cb5e-44b7-245ec0cf8f74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.866707] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935678, 'name': CreateVM_Task, 'duration_secs': 0.313475} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.866947] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 782.867711] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.007864] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.494s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.008460] env[61985]: DEBUG nova.compute.manager [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 783.011945] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.815s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.013659] env[61985]: INFO nova.compute.claims [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.106978] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5b15bfd9-ad08-412b-943d-2e2fbf9a40a0 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "4add2608-8187-42a2-9897-3e5a9049757e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.486s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.218834] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52928ae0-7c43-cb5e-44b7-245ec0cf8f74, 'name': SearchDatastore_Task, 'duration_secs': 0.009908} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.219127] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.223491] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.223596] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.225382] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.225382] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.225382] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.225382] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 783.225382] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8379d74e-8255-4b4d-88c2-918c10125a88 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.226749] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad8d65a3-44df-439a-9153-21f41f7cd588 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.238135] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 783.238135] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5230a6fc-f4a8-6814-217d-9f792655b61d" [ 783.238135] env[61985]: _type = "Task" [ 783.238135] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.239443] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.239613] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 783.243398] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeda7c87-c9e4-44ad-b76a-d7efba3e0390 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.254392] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5230a6fc-f4a8-6814-217d-9f792655b61d, 'name': SearchDatastore_Task, 'duration_secs': 0.008343} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.254767] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 783.254767] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521974a6-9174-efb8-5d5d-79650e51ad0e" [ 783.254767] env[61985]: _type = "Task" [ 783.254767] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.255455] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.255684] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.255890] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.267019] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521974a6-9174-efb8-5d5d-79650e51ad0e, 'name': SearchDatastore_Task, 'duration_secs': 0.008901} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.270337] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe950dae-8fa9-48b4-ad44-1a16e7f451cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.275978] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 783.275978] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521a0c88-5707-d997-3deb-a30cd0e4c322" [ 783.275978] env[61985]: _type = "Task" [ 783.275978] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.285782] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521a0c88-5707-d997-3deb-a30cd0e4c322, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.518901] env[61985]: DEBUG nova.compute.utils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.521225] env[61985]: DEBUG nova.compute.manager [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.521225] env[61985]: DEBUG nova.network.neutron [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 783.575585] env[61985]: DEBUG nova.network.neutron [req-6eb40481-5727-460c-834e-df9733df502e req-f6b6930a-3523-4a03-a70e-5c368d9416cc service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updated VIF entry in instance network info cache for port adde3d48-ebd4-450d-920a-fb93eeff4a8b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.581525] env[61985]: DEBUG nova.network.neutron [req-6eb40481-5727-460c-834e-df9733df502e req-f6b6930a-3523-4a03-a70e-5c368d9416cc service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance_info_cache with network_info: [{"id": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "address": "fa:16:3e:7b:a1:a8", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadde3d48-eb", "ovs_interfaceid": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.599908] env[61985]: DEBUG nova.policy [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '591e0f0c9ae94d0a8d2af1b8023016b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29145bcf8ccd41f6bf01ee34bd9b4d08', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 783.608028] env[61985]: DEBUG nova.network.neutron [req-83ef90e5-538e-4d7f-a9bb-90ad4e69abcb req-58f1fe26-7bc4-4966-8253-814301aec89d service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updated VIF entry in instance network info cache for port 4aa18b7b-89b5-4cad-af09-1bbc8f255029. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.608028] env[61985]: DEBUG nova.network.neutron [req-83ef90e5-538e-4d7f-a9bb-90ad4e69abcb req-58f1fe26-7bc4-4966-8253-814301aec89d service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance_info_cache with network_info: [{"id": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "address": "fa:16:3e:12:45:ef", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aa18b7b-89", "ovs_interfaceid": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.613881] env[61985]: DEBUG nova.compute.manager [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.790591] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521a0c88-5707-d997-3deb-a30cd0e4c322, 'name': SearchDatastore_Task, 'duration_secs': 0.022891} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.792265] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.792576] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 783.792862] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.793061] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.793272] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12fa92f3-2109-4022-8079-393e6d548700 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.797247] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ecfb4ebd-ee06-4b9c-8fb5-39b24e98bdfc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.804559] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 783.804559] env[61985]: value = "task-935679" [ 783.804559] env[61985]: _type = "Task" [ 783.804559] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.808895] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.810016] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 783.810105] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48f135e6-c8e4-4844-890c-47bf57528ea4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.815422] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935679, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.818295] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 783.818295] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523b4b22-1863-a293-45b5-cdcd0b62a2c1" [ 783.818295] env[61985]: _type = "Task" [ 783.818295] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.826304] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523b4b22-1863-a293-45b5-cdcd0b62a2c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.002605] env[61985]: DEBUG nova.network.neutron [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Successfully created port: cdd07618-1786-48b0-9f82-2fe473047b27 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 784.024422] env[61985]: DEBUG nova.compute.manager [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 784.083860] env[61985]: DEBUG oslo_concurrency.lockutils [req-6eb40481-5727-460c-834e-df9733df502e req-f6b6930a-3523-4a03-a70e-5c368d9416cc service nova] Releasing lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.115361] env[61985]: DEBUG oslo_concurrency.lockutils [req-83ef90e5-538e-4d7f-a9bb-90ad4e69abcb req-58f1fe26-7bc4-4966-8253-814301aec89d service nova] Releasing lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.137071] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.320709] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935679, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494903} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.323616] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 784.323781] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.326800] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1be35c7d-30af-45ae-b02c-b97762e3cfea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.335752] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523b4b22-1863-a293-45b5-cdcd0b62a2c1, 'name': SearchDatastore_Task, 'duration_secs': 0.008081} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.337733] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 784.337733] env[61985]: value = "task-935680" [ 784.337733] env[61985]: _type = "Task" [ 784.337733] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.338022] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d603a05-e1fb-4f27-bc74-ab61a86b9dac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.351080] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 784.351080] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ba1fc7-378e-2f87-8b35-ed136dd96cae" [ 784.351080] env[61985]: _type = "Task" [ 784.351080] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.354548] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935680, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.367880] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ba1fc7-378e-2f87-8b35-ed136dd96cae, 'name': SearchDatastore_Task, 'duration_secs': 0.010041} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.368058] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.368352] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] fea0cf39-e851-409f-86f5-31cc128a44dc/fea0cf39-e851-409f-86f5-31cc128a44dc.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 784.368628] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a607daa6-e7b3-4b8e-9a43-5ce227999ff4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.379892] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 784.379892] env[61985]: value = "task-935681" [ 784.379892] env[61985]: _type = "Task" [ 784.379892] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.388560] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935681, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.490084] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2caa9b58-294a-4c2c-93fb-ffeae3887e47 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.499417] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f07e8e3-c30a-48d2-987c-d004da0aeaee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.531665] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca12e0a4-2f59-4655-b60c-665c9c5a9477 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.543222] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1c341c-3b88-4fb5-b27a-87a1bae3a6bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.557924] env[61985]: DEBUG nova.compute.provider_tree [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.850558] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935680, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.421862} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.851172] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.852035] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6f5be1-e3ef-4df9-8077-6babec1682b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.874721] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.876179] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4296b631-e802-4379-8199-2d2f8d90de5d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.898893] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935681, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488517} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.900145] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] fea0cf39-e851-409f-86f5-31cc128a44dc/fea0cf39-e851-409f-86f5-31cc128a44dc.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 784.900394] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.900694] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 784.900694] env[61985]: value = "task-935682" [ 784.900694] env[61985]: _type = "Task" [ 784.900694] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.900874] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da12d775-a61f-48e9-a82b-ef698c043960 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.910383] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935682, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.911537] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 784.911537] env[61985]: value = "task-935683" [ 784.911537] env[61985]: _type = "Task" [ 784.911537] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.919394] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935683, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.040521] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquiring lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.040706] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.043908] env[61985]: DEBUG nova.compute.manager [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 785.061860] env[61985]: DEBUG nova.scheduler.client.report [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.085308] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.085571] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.085728] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.085910] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.086072] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.086467] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.086467] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.086604] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.086800] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.086916] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.087118] env[61985]: DEBUG nova.virt.hardware [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.088361] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ab0ace-7146-4d1b-8f37-e1ade31b3b72 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.099112] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb886b0-5156-4f3a-b7c6-fb5bc035fbdb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.413812] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935682, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.421682] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935683, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.264602} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.422137] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.423026] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed00df79-efba-4bb3-b0ed-e4624b3cbb8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.446055] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] fea0cf39-e851-409f-86f5-31cc128a44dc/fea0cf39-e851-409f-86f5-31cc128a44dc.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.447882] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac2114b8-dd65-418f-96c1-19f65f3bccd5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.466734] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 785.466734] env[61985]: value = "task-935684" [ 785.466734] env[61985]: _type = "Task" [ 785.466734] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.474792] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935684, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.575028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.575028] env[61985]: DEBUG nova.compute.manager [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 785.577730] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.327s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.579708] env[61985]: INFO nova.compute.claims [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.780305] env[61985]: DEBUG nova.network.neutron [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Successfully updated port: cdd07618-1786-48b0-9f82-2fe473047b27 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 785.790868] env[61985]: DEBUG nova.compute.manager [req-6dcde5c7-1f02-4cbf-b7a6-836642c7f5d6 req-9100c1ef-0709-4d21-b335-86a0fcead295 service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Received event network-vif-plugged-cdd07618-1786-48b0-9f82-2fe473047b27 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 785.790868] env[61985]: DEBUG oslo_concurrency.lockutils [req-6dcde5c7-1f02-4cbf-b7a6-836642c7f5d6 req-9100c1ef-0709-4d21-b335-86a0fcead295 service nova] Acquiring lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.790868] env[61985]: DEBUG oslo_concurrency.lockutils [req-6dcde5c7-1f02-4cbf-b7a6-836642c7f5d6 req-9100c1ef-0709-4d21-b335-86a0fcead295 service nova] Lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.790868] env[61985]: DEBUG oslo_concurrency.lockutils [req-6dcde5c7-1f02-4cbf-b7a6-836642c7f5d6 req-9100c1ef-0709-4d21-b335-86a0fcead295 service nova] Lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.790868] env[61985]: DEBUG nova.compute.manager [req-6dcde5c7-1f02-4cbf-b7a6-836642c7f5d6 req-9100c1ef-0709-4d21-b335-86a0fcead295 service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] No waiting events found dispatching network-vif-plugged-cdd07618-1786-48b0-9f82-2fe473047b27 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 785.791213] env[61985]: WARNING nova.compute.manager [req-6dcde5c7-1f02-4cbf-b7a6-836642c7f5d6 req-9100c1ef-0709-4d21-b335-86a0fcead295 service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Received unexpected event network-vif-plugged-cdd07618-1786-48b0-9f82-2fe473047b27 for instance with vm_state building and task_state spawning. [ 785.913233] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935682, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.976607] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935684, 'name': ReconfigVM_Task, 'duration_secs': 0.274801} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.976899] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Reconfigured VM instance instance-00000027 to attach disk [datastore2] fea0cf39-e851-409f-86f5-31cc128a44dc/fea0cf39-e851-409f-86f5-31cc128a44dc.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 785.977633] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d106a394-de62-42f4-81eb-c7b355102d90 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.983642] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 785.983642] env[61985]: value = "task-935685" [ 785.983642] env[61985]: _type = "Task" [ 785.983642] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.991571] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935685, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.086199] env[61985]: DEBUG nova.compute.utils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.089792] env[61985]: DEBUG nova.compute.manager [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.089792] env[61985]: DEBUG nova.network.neutron [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 786.143054] env[61985]: DEBUG nova.policy [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'baf9793f2d8a41698e11196f639fd749', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd49f3925a4074d779d6dd68164545cfb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 786.284132] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquiring lock "refresh_cache-3815af53-ba3b-4c60-a5fd-2d94102cdd42" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.284330] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquired lock "refresh_cache-3815af53-ba3b-4c60-a5fd-2d94102cdd42" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.284469] env[61985]: DEBUG nova.network.neutron [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 786.316790] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "4add2608-8187-42a2-9897-3e5a9049757e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.317059] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "4add2608-8187-42a2-9897-3e5a9049757e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.317279] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "4add2608-8187-42a2-9897-3e5a9049757e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.317467] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "4add2608-8187-42a2-9897-3e5a9049757e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.317715] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "4add2608-8187-42a2-9897-3e5a9049757e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.319849] env[61985]: INFO nova.compute.manager [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Terminating instance [ 786.321572] env[61985]: DEBUG nova.compute.manager [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 786.321776] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 786.322660] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1858cc-6015-47a9-8c69-09e5f1805575 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.330485] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 786.331253] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32a3e992-4746-42dd-93b9-75a0a7583295 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.336924] env[61985]: DEBUG oslo_vmware.api [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for the task: (returnval){ [ 786.336924] env[61985]: value = "task-935686" [ 786.336924] env[61985]: _type = "Task" [ 786.336924] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.344853] env[61985]: DEBUG oslo_vmware.api [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935686, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.416254] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935682, 'name': ReconfigVM_Task, 'duration_secs': 1.127678} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.417237] env[61985]: DEBUG nova.network.neutron [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Successfully created port: 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.420233] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.420233] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68ec73fc-ba14-4c26-8048-c3c085ec7acf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.429017] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 786.429017] env[61985]: value = "task-935687" [ 786.429017] env[61985]: _type = "Task" [ 786.429017] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.435485] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935687, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.494884] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935685, 'name': Rename_Task, 'duration_secs': 0.141512} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.496894] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 786.496894] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-471a1003-3e39-4d09-a499-a25543c016aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.501793] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 786.501793] env[61985]: value = "task-935688" [ 786.501793] env[61985]: _type = "Task" [ 786.501793] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.509851] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935688, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.598397] env[61985]: DEBUG nova.compute.manager [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 786.830120] env[61985]: DEBUG nova.network.neutron [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.847259] env[61985]: DEBUG oslo_vmware.api [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935686, 'name': PowerOffVM_Task, 'duration_secs': 0.257214} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.851550] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 786.851748] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 786.852204] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c94f537a-30e5-4163-9511-78d717829fb2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.937913] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935687, 'name': Rename_Task, 'duration_secs': 0.190198} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.938215] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 786.938469] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cb28c22-aa2a-4a61-8f0d-5b2d056b9b2c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.945126] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 786.945126] env[61985]: value = "task-935690" [ 786.945126] env[61985]: _type = "Task" [ 786.945126] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.957460] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935690, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.966792] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 786.967018] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 786.967229] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Deleting the datastore file [datastore1] 4add2608-8187-42a2-9897-3e5a9049757e {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.967452] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd9d52a1-8dd4-46b6-a8cf-da70b1ca11ac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.973871] env[61985]: DEBUG oslo_vmware.api [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for the task: (returnval){ [ 786.973871] env[61985]: value = "task-935691" [ 786.973871] env[61985]: _type = "Task" [ 786.973871] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.981718] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba2b44a-31f1-4f66-8d73-7c7d8bccd01c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.991872] env[61985]: DEBUG oslo_vmware.api [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935691, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.994793] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d3428f-3644-4588-b50e-0d33b860cbd3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.033566] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007d27f9-358e-4d07-bcaf-d02005354fb4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.044266] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94ebe01-6a88-41a7-b158-6ccff9f779fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.048019] env[61985]: DEBUG oslo_vmware.api [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935688, 'name': PowerOnVM_Task, 'duration_secs': 0.495044} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.048333] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 787.048548] env[61985]: INFO nova.compute.manager [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Took 8.00 seconds to spawn the instance on the hypervisor. [ 787.048784] env[61985]: DEBUG nova.compute.manager [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 787.049926] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d9a386-384d-4238-ae5f-ca952e00a29f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.061576] env[61985]: DEBUG nova.compute.provider_tree [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.079862] env[61985]: DEBUG nova.network.neutron [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Updating instance_info_cache with network_info: [{"id": "cdd07618-1786-48b0-9f82-2fe473047b27", "address": "fa:16:3e:4d:ed:9f", "network": {"id": "807bff2f-2e49-4f1a-9c28-c55667e9a55e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1780060480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "29145bcf8ccd41f6bf01ee34bd9b4d08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdd07618-17", "ovs_interfaceid": "cdd07618-1786-48b0-9f82-2fe473047b27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.457943] env[61985]: DEBUG oslo_vmware.api [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935690, 'name': PowerOnVM_Task, 'duration_secs': 0.459723} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.457943] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 787.457943] env[61985]: INFO nova.compute.manager [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Took 5.91 seconds to spawn the instance on the hypervisor. [ 787.457943] env[61985]: DEBUG nova.compute.manager [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 787.457943] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02530034-ecf5-4168-9f1c-1da8078d8a50 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.482843] env[61985]: DEBUG oslo_vmware.api [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Task: {'id': task-935691, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245681} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.483110] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.483296] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 787.483472] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 787.483644] env[61985]: INFO nova.compute.manager [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Took 1.16 seconds to destroy the instance on the hypervisor. [ 787.483911] env[61985]: DEBUG oslo.service.loopingcall [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.484091] env[61985]: DEBUG nova.compute.manager [-] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.484200] env[61985]: DEBUG nova.network.neutron [-] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 787.566191] env[61985]: DEBUG nova.scheduler.client.report [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.577775] env[61985]: INFO nova.compute.manager [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Took 29.32 seconds to build instance. [ 787.581792] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Releasing lock "refresh_cache-3815af53-ba3b-4c60-a5fd-2d94102cdd42" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.585015] env[61985]: DEBUG nova.compute.manager [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Instance network_info: |[{"id": "cdd07618-1786-48b0-9f82-2fe473047b27", "address": "fa:16:3e:4d:ed:9f", "network": {"id": "807bff2f-2e49-4f1a-9c28-c55667e9a55e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1780060480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "29145bcf8ccd41f6bf01ee34bd9b4d08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdd07618-17", "ovs_interfaceid": "cdd07618-1786-48b0-9f82-2fe473047b27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 787.585488] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:ed:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ed91b7b-b4ec-486d-ab34-af0afb7ec691', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cdd07618-1786-48b0-9f82-2fe473047b27', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.593477] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Creating folder: Project (29145bcf8ccd41f6bf01ee34bd9b4d08). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 787.595602] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8be27c3d-87af-4956-bad7-31f30ad5698c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.606942] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Created folder: Project (29145bcf8ccd41f6bf01ee34bd9b4d08) in parent group-v211285. [ 787.607582] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Creating folder: Instances. Parent ref: group-v211310. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 787.607954] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21da59a0-f916-44da-8899-00a695f4ce09 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.610558] env[61985]: DEBUG nova.compute.manager [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 787.624019] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Created folder: Instances in parent group-v211310. [ 787.624019] env[61985]: DEBUG oslo.service.loopingcall [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.624019] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 787.624019] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3ae9768-b230-4ca2-bde4-303c7f3013dc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.649178] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.649178] env[61985]: value = "task-935694" [ 787.649178] env[61985]: _type = "Task" [ 787.649178] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.653103] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.653540] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.653822] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.654147] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.654495] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.655039] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.655549] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.655820] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.656112] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.656379] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.656649] env[61985]: DEBUG nova.virt.hardware [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.657540] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6d1281-7437-42c0-bfca-721d4e585efc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.668399] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935694, 'name': CreateVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.670519] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f523906c-6165-4c31-90ff-d0223c268333 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.898597] env[61985]: DEBUG nova.compute.manager [req-d221732f-4972-4530-ae21-e7efd45cfef3 req-8163cd3d-0658-4fce-83c5-da8d9adf079c service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Received event network-changed-cdd07618-1786-48b0-9f82-2fe473047b27 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 787.898597] env[61985]: DEBUG nova.compute.manager [req-d221732f-4972-4530-ae21-e7efd45cfef3 req-8163cd3d-0658-4fce-83c5-da8d9adf079c service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Refreshing instance network info cache due to event network-changed-cdd07618-1786-48b0-9f82-2fe473047b27. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 787.898597] env[61985]: DEBUG oslo_concurrency.lockutils [req-d221732f-4972-4530-ae21-e7efd45cfef3 req-8163cd3d-0658-4fce-83c5-da8d9adf079c service nova] Acquiring lock "refresh_cache-3815af53-ba3b-4c60-a5fd-2d94102cdd42" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.898597] env[61985]: DEBUG oslo_concurrency.lockutils [req-d221732f-4972-4530-ae21-e7efd45cfef3 req-8163cd3d-0658-4fce-83c5-da8d9adf079c service nova] Acquired lock "refresh_cache-3815af53-ba3b-4c60-a5fd-2d94102cdd42" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.898597] env[61985]: DEBUG nova.network.neutron [req-d221732f-4972-4530-ae21-e7efd45cfef3 req-8163cd3d-0658-4fce-83c5-da8d9adf079c service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Refreshing network info cache for port cdd07618-1786-48b0-9f82-2fe473047b27 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 787.977924] env[61985]: INFO nova.compute.manager [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Took 27.82 seconds to build instance. [ 788.070906] env[61985]: DEBUG nova.network.neutron [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Successfully updated port: 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 788.070906] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.071589] env[61985]: DEBUG nova.compute.manager [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 788.078716] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.794s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.081940] env[61985]: INFO nova.compute.claims [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.086949] env[61985]: DEBUG nova.compute.manager [req-b9fa9910-91a7-41f1-abb0-2d1dc02eb023 req-97d82c16-3555-4c89-b58f-6a061ef3d106 service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Received event network-vif-deleted-4afe5581-970d-4d14-b68d-76a76308741e {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 788.087264] env[61985]: INFO nova.compute.manager [req-b9fa9910-91a7-41f1-abb0-2d1dc02eb023 req-97d82c16-3555-4c89-b58f-6a061ef3d106 service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Neutron deleted interface 4afe5581-970d-4d14-b68d-76a76308741e; detaching it from the instance and deleting it from the info cache [ 788.087513] env[61985]: DEBUG nova.network.neutron [req-b9fa9910-91a7-41f1-abb0-2d1dc02eb023 req-97d82c16-3555-4c89-b58f-6a061ef3d106 service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.090153] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fba3147-5076-4d28-9ee6-d2f09807f8b5 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.598s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.159266] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935694, 'name': CreateVM_Task, 'duration_secs': 0.498003} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.163799] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 788.164856] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.165031] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.165400] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 788.165654] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12b2a294-37ed-4a71-b558-a8f0dc731a9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.177754] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for the task: (returnval){ [ 788.177754] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5228a0ac-53ed-53fe-1e67-72f2fa4c3573" [ 788.177754] env[61985]: _type = "Task" [ 788.177754] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.189160] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5228a0ac-53ed-53fe-1e67-72f2fa4c3573, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.385606] env[61985]: DEBUG nova.network.neutron [-] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.480698] env[61985]: DEBUG oslo_concurrency.lockutils [None req-cabed679-0c03-4e4b-9bb8-186124f9606f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "794cff68-6e26-4607-96f7-eaeb41182551" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.672s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.579224] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.579519] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquired lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.579756] env[61985]: DEBUG nova.network.neutron [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 788.592115] env[61985]: DEBUG nova.compute.utils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.600373] env[61985]: DEBUG nova.compute.manager [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 788.600487] env[61985]: DEBUG nova.network.neutron [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 788.603410] env[61985]: DEBUG nova.compute.manager [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.611852] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-375b1460-0502-465a-9556-7656fd7d5474 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.623314] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e5d2c5-5b17-4efc-834b-7a5861f39925 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.659855] env[61985]: DEBUG nova.compute.manager [req-b9fa9910-91a7-41f1-abb0-2d1dc02eb023 req-97d82c16-3555-4c89-b58f-6a061ef3d106 service nova] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Detach interface failed, port_id=4afe5581-970d-4d14-b68d-76a76308741e, reason: Instance 4add2608-8187-42a2-9897-3e5a9049757e could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 788.665115] env[61985]: INFO nova.compute.manager [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Rebuilding instance [ 788.690226] env[61985]: DEBUG nova.policy [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '344f7f484f4240cd9dc32b200b697e12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfc57e6393c94545be32165d41230db3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 788.698820] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5228a0ac-53ed-53fe-1e67-72f2fa4c3573, 'name': SearchDatastore_Task, 'duration_secs': 0.022911} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.699360] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.699617] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.699919] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.700122] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.700347] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.700634] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4a6b2d3-9697-4be2-8e20-b1f093e726e4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.710218] env[61985]: DEBUG nova.network.neutron [req-d221732f-4972-4530-ae21-e7efd45cfef3 req-8163cd3d-0658-4fce-83c5-da8d9adf079c service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Updated VIF entry in instance network info cache for port cdd07618-1786-48b0-9f82-2fe473047b27. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.710638] env[61985]: DEBUG nova.network.neutron [req-d221732f-4972-4530-ae21-e7efd45cfef3 req-8163cd3d-0658-4fce-83c5-da8d9adf079c service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Updating instance_info_cache with network_info: [{"id": "cdd07618-1786-48b0-9f82-2fe473047b27", "address": "fa:16:3e:4d:ed:9f", "network": {"id": "807bff2f-2e49-4f1a-9c28-c55667e9a55e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1780060480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "29145bcf8ccd41f6bf01ee34bd9b4d08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdd07618-17", "ovs_interfaceid": "cdd07618-1786-48b0-9f82-2fe473047b27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.720497] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.720608] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 788.723641] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55b73cbf-ea6c-4c6e-8a4f-6f5ea59d6450 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.733058] env[61985]: DEBUG nova.compute.manager [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.734117] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60699fc5-9509-4c00-a824-97ef3ab4d6b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.738412] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for the task: (returnval){ [ 788.738412] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529dd25b-d61e-afa6-ace4-b15a90fb63fb" [ 788.738412] env[61985]: _type = "Task" [ 788.738412] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.753409] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529dd25b-d61e-afa6-ace4-b15a90fb63fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.890449] env[61985]: INFO nova.compute.manager [-] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Took 1.41 seconds to deallocate network for instance. [ 788.986374] env[61985]: DEBUG nova.compute.manager [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 789.055750] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6458a1-2fa4-4e84-b26b-6e594020c346 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.063790] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e84b11-1454-4fd4-9064-10f9ff762930 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.098933] env[61985]: DEBUG nova.network.neutron [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Successfully created port: 787b8b7e-dea3-41b7-9349-4f87afe7b88d {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.100905] env[61985]: DEBUG nova.compute.manager [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 789.103904] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4324e769-acde-4ce9-8877-0c6603aec978 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.115733] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da22c9ca-6478-4f95-8692-c9fde133e64e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.132728] env[61985]: DEBUG nova.compute.provider_tree [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 789.138020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.138020] env[61985]: DEBUG nova.network.neutron [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.146174] env[61985]: DEBUG nova.compute.manager [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Stashing vm_state: active {{(pid=61985) _prep_resize /opt/stack/nova/nova/compute/manager.py:5630}} [ 789.213428] env[61985]: DEBUG oslo_concurrency.lockutils [req-d221732f-4972-4530-ae21-e7efd45cfef3 req-8163cd3d-0658-4fce-83c5-da8d9adf079c service nova] Releasing lock "refresh_cache-3815af53-ba3b-4c60-a5fd-2d94102cdd42" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.249965] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 789.250205] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529dd25b-d61e-afa6-ace4-b15a90fb63fb, 'name': SearchDatastore_Task, 'duration_secs': 0.027927} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.250410] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63a27078-54ab-405e-a76c-b7b4014c20fd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.252392] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac42ba4f-55fa-4440-a79a-41a0df595ac2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.257345] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for the task: (returnval){ [ 789.257345] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5229dc40-f8ab-728a-352b-275e41e98377" [ 789.257345] env[61985]: _type = "Task" [ 789.257345] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.261349] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 789.261349] env[61985]: value = "task-935695" [ 789.261349] env[61985]: _type = "Task" [ 789.261349] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.267908] env[61985]: DEBUG nova.network.neutron [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updating instance_info_cache with network_info: [{"id": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "address": "fa:16:3e:d4:ac:00", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60c2bc37-31", "ovs_interfaceid": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.269099] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5229dc40-f8ab-728a-352b-275e41e98377, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.274387] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935695, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.397113] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.510499] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.659161] env[61985]: ERROR nova.scheduler.client.report [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [req-e186b4fa-b3cb-436a-82ef-68f631d02eb6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e186b4fa-b3cb-436a-82ef-68f631d02eb6"}]} [ 789.671757] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.675366] env[61985]: DEBUG nova.scheduler.client.report [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 789.691116] env[61985]: DEBUG nova.scheduler.client.report [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 789.691343] env[61985]: DEBUG nova.compute.provider_tree [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 789.703296] env[61985]: DEBUG nova.scheduler.client.report [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 789.721404] env[61985]: DEBUG nova.scheduler.client.report [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 789.770754] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Releasing lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.771022] env[61985]: DEBUG nova.compute.manager [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Instance network_info: |[{"id": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "address": "fa:16:3e:d4:ac:00", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60c2bc37-31", "ovs_interfaceid": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 789.771318] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5229dc40-f8ab-728a-352b-275e41e98377, 'name': SearchDatastore_Task, 'duration_secs': 0.00963} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.772674] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:ac:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '60c2bc37-3190-4dfb-8b71-fd6eb3c3949b', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.779569] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Creating folder: Project (d49f3925a4074d779d6dd68164545cfb). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 789.779865] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.780234] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 3815af53-ba3b-4c60-a5fd-2d94102cdd42/3815af53-ba3b-4c60-a5fd-2d94102cdd42.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 789.783692] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-855f0ba2-9603-4585-9dc5-034ee43f26eb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.785296] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dae372a5-ba7e-4b4e-9648-029498ce1c39 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.787518] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935695, 'name': PowerOffVM_Task, 'duration_secs': 0.143533} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.789987] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 789.790232] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 789.791721] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2dee54-8a50-4a4e-819e-8425e8fb6538 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.796313] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for the task: (returnval){ [ 789.796313] env[61985]: value = "task-935697" [ 789.796313] env[61985]: _type = "Task" [ 789.796313] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.806016] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Created folder: Project (d49f3925a4074d779d6dd68164545cfb) in parent group-v211285. [ 789.806016] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Creating folder: Instances. Parent ref: group-v211313. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 789.806379] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4bf1ea56-d3c0-428b-8f3e-b2c62158675e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.811573] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935697, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.814770] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 789.815027] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a202fef-f4af-42e5-a47a-d283f60e103c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.820039] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Created folder: Instances in parent group-v211313. [ 789.820276] env[61985]: DEBUG oslo.service.loopingcall [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.820488] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 789.820648] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a440c04-d04f-47af-920c-61c067654f60 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.840707] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 789.840707] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 789.840944] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Deleting the datastore file [datastore2] 794cff68-6e26-4607-96f7-eaeb41182551 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 789.841172] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2296daa4-5d1b-4104-9645-3bcc12888e78 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.846022] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.846022] env[61985]: value = "task-935700" [ 789.846022] env[61985]: _type = "Task" [ 789.846022] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.848386] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 789.848386] env[61985]: value = "task-935701" [ 789.848386] env[61985]: _type = "Task" [ 789.848386] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.859326] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935700, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.860035] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935701, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.948285] env[61985]: DEBUG nova.compute.manager [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Received event network-vif-plugged-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 789.948582] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] Acquiring lock "3694e20c-ce37-4097-9991-8a06f38b2734-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.948714] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] Lock "3694e20c-ce37-4097-9991-8a06f38b2734-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.948883] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] Lock "3694e20c-ce37-4097-9991-8a06f38b2734-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.949067] env[61985]: DEBUG nova.compute.manager [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] No waiting events found dispatching network-vif-plugged-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 789.949233] env[61985]: WARNING nova.compute.manager [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Received unexpected event network-vif-plugged-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b for instance with vm_state building and task_state spawning. [ 789.949390] env[61985]: DEBUG nova.compute.manager [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Received event network-changed-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 789.949540] env[61985]: DEBUG nova.compute.manager [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Refreshing instance network info cache due to event network-changed-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 789.949720] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] Acquiring lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.949858] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] Acquired lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.950015] env[61985]: DEBUG nova.network.neutron [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Refreshing network info cache for port 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 790.114834] env[61985]: DEBUG nova.compute.manager [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 790.130841] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4f6306-0aed-4f45-ab16-7d596de76328 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.143789] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.144015] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.144185] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.144377] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.144549] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.144720] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.144998] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.145214] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.145451] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.145621] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.145749] env[61985]: DEBUG nova.virt.hardware [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.147018] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15233ad7-8d9e-498b-8f0a-500f9134f8c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.152284] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cadd3b2-99a1-4788-8a0d-b903973e55fd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.190684] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616eb74a-53d1-451c-8bab-bb2ac9781b68 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.197325] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444b29da-8375-4aed-b19c-d1aebdf36e2e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.213652] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb3bba8-76dd-4207-a2d3-a7ac13fe7c1b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.231449] env[61985]: DEBUG nova.compute.provider_tree [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 790.306610] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935697, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.355767] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935700, 'name': CreateVM_Task, 'duration_secs': 0.373491} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.356281] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 790.356982] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.357183] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.357538] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 790.360472] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24afede7-6926-4a17-8329-0678a9f2d0cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.362091] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935701, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09546} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.362342] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.362527] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 790.362699] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 790.367564] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 790.367564] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529e99ab-de22-092f-921b-1b0eb9765be6" [ 790.367564] env[61985]: _type = "Task" [ 790.367564] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.375272] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529e99ab-de22-092f-921b-1b0eb9765be6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.530034] env[61985]: DEBUG nova.compute.manager [req-b686fd0e-fc5a-4f22-83c3-867e945752ef req-3055514b-0909-4326-a2a4-1858d29cbf7d service nova] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Received event network-vif-plugged-787b8b7e-dea3-41b7-9349-4f87afe7b88d {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 790.530272] env[61985]: DEBUG oslo_concurrency.lockutils [req-b686fd0e-fc5a-4f22-83c3-867e945752ef req-3055514b-0909-4326-a2a4-1858d29cbf7d service nova] Acquiring lock "f8bc8b71-0317-479d-b2f9-9471757f0774-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.530475] env[61985]: DEBUG oslo_concurrency.lockutils [req-b686fd0e-fc5a-4f22-83c3-867e945752ef req-3055514b-0909-4326-a2a4-1858d29cbf7d service nova] Lock "f8bc8b71-0317-479d-b2f9-9471757f0774-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.530639] env[61985]: DEBUG oslo_concurrency.lockutils [req-b686fd0e-fc5a-4f22-83c3-867e945752ef req-3055514b-0909-4326-a2a4-1858d29cbf7d service nova] Lock "f8bc8b71-0317-479d-b2f9-9471757f0774-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.530808] env[61985]: DEBUG nova.compute.manager [req-b686fd0e-fc5a-4f22-83c3-867e945752ef req-3055514b-0909-4326-a2a4-1858d29cbf7d service nova] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] No waiting events found dispatching network-vif-plugged-787b8b7e-dea3-41b7-9349-4f87afe7b88d {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 790.530979] env[61985]: WARNING nova.compute.manager [req-b686fd0e-fc5a-4f22-83c3-867e945752ef req-3055514b-0909-4326-a2a4-1858d29cbf7d service nova] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Received unexpected event network-vif-plugged-787b8b7e-dea3-41b7-9349-4f87afe7b88d for instance with vm_state building and task_state spawning. [ 790.645364] env[61985]: DEBUG nova.network.neutron [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updated VIF entry in instance network info cache for port 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.645781] env[61985]: DEBUG nova.network.neutron [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updating instance_info_cache with network_info: [{"id": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "address": "fa:16:3e:d4:ac:00", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60c2bc37-31", "ovs_interfaceid": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.748040] env[61985]: DEBUG nova.network.neutron [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Successfully updated port: 787b8b7e-dea3-41b7-9349-4f87afe7b88d {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.773177] env[61985]: DEBUG nova.scheduler.client.report [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 73 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 790.773402] env[61985]: DEBUG nova.compute.provider_tree [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 73 to 74 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 790.773556] env[61985]: DEBUG nova.compute.provider_tree [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 44, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 790.808257] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935697, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582862} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.808504] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 3815af53-ba3b-4c60-a5fd-2d94102cdd42/3815af53-ba3b-4c60-a5fd-2d94102cdd42.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 790.808720] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 790.808963] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39642557-398e-4ba9-97cd-967d0637e509 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.816514] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for the task: (returnval){ [ 790.816514] env[61985]: value = "task-935702" [ 790.816514] env[61985]: _type = "Task" [ 790.816514] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.826311] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935702, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.881808] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529e99ab-de22-092f-921b-1b0eb9765be6, 'name': SearchDatastore_Task, 'duration_secs': 0.03103} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.881808] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.881808] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.882040] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.882040] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.882194] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.882463] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fae742fe-0fe0-442f-a52d-986efa6061cd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.899176] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.899176] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 790.899176] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4607b6f-678e-44e8-b240-6a78d1958f0e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.904612] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 790.904612] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52753e7b-532c-28dd-df60-49e1bd93a055" [ 790.904612] env[61985]: _type = "Task" [ 790.904612] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.914793] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52753e7b-532c-28dd-df60-49e1bd93a055, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.148724] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7a71456-e1e0-4db9-91b9-45fe6cf0e2fd req-0803cf39-ae64-49f9-8260-2812e42dd139 service nova] Releasing lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.253327] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-f8bc8b71-0317-479d-b2f9-9471757f0774" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.253483] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-f8bc8b71-0317-479d-b2f9-9471757f0774" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.253659] env[61985]: DEBUG nova.network.neutron [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.278556] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.201s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.279031] env[61985]: DEBUG nova.compute.manager [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 791.281385] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.842s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.282725] env[61985]: INFO nova.compute.claims [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.325962] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935702, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065862} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.326241] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 791.326992] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8defcc-8288-4f16-81d8-d5afb1fe192a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.348921] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 3815af53-ba3b-4c60-a5fd-2d94102cdd42/3815af53-ba3b-4c60-a5fd-2d94102cdd42.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 791.349179] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-608a943a-1570-40b9-989c-1f902790ed05 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.369285] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for the task: (returnval){ [ 791.369285] env[61985]: value = "task-935703" [ 791.369285] env[61985]: _type = "Task" [ 791.369285] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.380395] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935703, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.397570] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 791.397852] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 791.398027] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.398469] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 791.398469] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.398624] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 791.398851] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 791.399028] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 791.399206] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 791.399371] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 791.399581] env[61985]: DEBUG nova.virt.hardware [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 791.400677] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6b2b95-a505-4bb2-baeb-33d523370416 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.410882] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef3557f-d7c1-4c6e-9ddc-f84b1859afe9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.421982] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52753e7b-532c-28dd-df60-49e1bd93a055, 'name': SearchDatastore_Task, 'duration_secs': 0.082937} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.429572] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 791.435311] env[61985]: DEBUG oslo.service.loopingcall [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.435553] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee90c990-0d7e-4600-88b1-8e4f0e61de12 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.437694] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 791.437904] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28d0af24-5213-4d2f-9cca-5059539f69fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.452352] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 791.452352] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52295acb-d58a-16da-d67e-8400bc699ed4" [ 791.452352] env[61985]: _type = "Task" [ 791.452352] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.456037] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 791.456037] env[61985]: value = "task-935704" [ 791.456037] env[61985]: _type = "Task" [ 791.456037] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.461770] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52295acb-d58a-16da-d67e-8400bc699ed4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.466641] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935704, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.787294] env[61985]: DEBUG nova.compute.utils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 791.790524] env[61985]: DEBUG nova.compute.manager [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 791.790694] env[61985]: DEBUG nova.network.neutron [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 791.810910] env[61985]: DEBUG nova.network.neutron [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.837140] env[61985]: DEBUG nova.policy [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26d6f86b74ef4f5ca7af4c94176bdf35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e8e226d5c9240928ccc35013189235c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 791.879606] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935703, 'name': ReconfigVM_Task, 'duration_secs': 0.327552} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.883161] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 3815af53-ba3b-4c60-a5fd-2d94102cdd42/3815af53-ba3b-4c60-a5fd-2d94102cdd42.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.884279] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-26169a09-a2b4-40a5-a667-8e0dc625adf8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.893022] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for the task: (returnval){ [ 791.893022] env[61985]: value = "task-935705" [ 791.893022] env[61985]: _type = "Task" [ 791.893022] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.901560] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935705, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.967182] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52295acb-d58a-16da-d67e-8400bc699ed4, 'name': SearchDatastore_Task, 'duration_secs': 0.019023} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.970936] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.971344] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 3694e20c-ce37-4097-9991-8a06f38b2734/3694e20c-ce37-4097-9991-8a06f38b2734.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 791.971643] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935704, 'name': CreateVM_Task, 'duration_secs': 0.247559} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.971896] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af2985ca-b6cc-4737-9fa3-fd641a0c671a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.974526] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 791.974967] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.975148] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.975606] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.976225] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec5e03c9-5695-46b8-9431-5c57951e9aff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.979644] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 791.979644] env[61985]: value = "task-935706" [ 791.979644] env[61985]: _type = "Task" [ 791.979644] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.980850] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 791.980850] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529a3fc9-76ce-cef2-43fb-2c877bd2e56b" [ 791.980850] env[61985]: _type = "Task" [ 791.980850] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.991913] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935706, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.994734] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529a3fc9-76ce-cef2-43fb-2c877bd2e56b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.057114] env[61985]: DEBUG nova.network.neutron [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Updating instance_info_cache with network_info: [{"id": "787b8b7e-dea3-41b7-9349-4f87afe7b88d", "address": "fa:16:3e:2a:b2:dc", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap787b8b7e-de", "ovs_interfaceid": "787b8b7e-dea3-41b7-9349-4f87afe7b88d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.109291] env[61985]: DEBUG nova.network.neutron [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Successfully created port: bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 792.292431] env[61985]: DEBUG nova.compute.manager [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 792.402889] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935705, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.494796] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935706, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.499164] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529a3fc9-76ce-cef2-43fb-2c877bd2e56b, 'name': SearchDatastore_Task, 'duration_secs': 0.013645} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.499435] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.499687] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 792.499976] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.500149] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.500315] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 792.502581] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a031c464-6a81-4cff-98b7-f03cef394409 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.515655] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 792.515836] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 792.516545] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6eaf35cb-c6da-418c-9114-12db13d8a92b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.521093] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 792.521093] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5247df8e-ee55-5ecb-c83d-28e50276fe10" [ 792.521093] env[61985]: _type = "Task" [ 792.521093] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.531833] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5247df8e-ee55-5ecb-c83d-28e50276fe10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.562009] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-f8bc8b71-0317-479d-b2f9-9471757f0774" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.562579] env[61985]: DEBUG nova.compute.manager [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Instance network_info: |[{"id": "787b8b7e-dea3-41b7-9349-4f87afe7b88d", "address": "fa:16:3e:2a:b2:dc", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap787b8b7e-de", "ovs_interfaceid": "787b8b7e-dea3-41b7-9349-4f87afe7b88d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 792.563555] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:b2:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bff6c3a1-cc80-46ca-86c0-6dbb029edddb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '787b8b7e-dea3-41b7-9349-4f87afe7b88d', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.573529] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Creating folder: Project (cfc57e6393c94545be32165d41230db3). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.578942] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2025d5ca-6b8b-45e3-883e-7c5934aa91d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.589363] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Created folder: Project (cfc57e6393c94545be32165d41230db3) in parent group-v211285. [ 792.589654] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Creating folder: Instances. Parent ref: group-v211317. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.589810] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2eb1c13e-0a5e-456b-9740-9348f18e990d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.603419] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Created folder: Instances in parent group-v211317. [ 792.603673] env[61985]: DEBUG oslo.service.loopingcall [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.604354] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 792.604667] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49c2d530-4be7-4dce-a468-c06e63b90da3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.629718] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.629718] env[61985]: value = "task-935709" [ 792.629718] env[61985]: _type = "Task" [ 792.629718] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.642188] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935709, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.677157] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3526a92b-e275-434e-b464-ecf6ce7e9a95 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.686283] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b07b762-65f1-46b8-97d3-da4b403ca6e5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.735445] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be091995-565d-4ed0-ade3-bf9d35b642a5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.743708] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5d661d-b3d7-41a4-b568-e9acb3d70b6b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.758842] env[61985]: DEBUG nova.compute.provider_tree [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 792.781621] env[61985]: DEBUG nova.compute.manager [req-aa2c0d3b-246c-4849-984d-72a11d08affa req-23aa2244-b230-4431-8486-2fa4a7e5fbb3 service nova] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Received event network-changed-787b8b7e-dea3-41b7-9349-4f87afe7b88d {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 792.781900] env[61985]: DEBUG nova.compute.manager [req-aa2c0d3b-246c-4849-984d-72a11d08affa req-23aa2244-b230-4431-8486-2fa4a7e5fbb3 service nova] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Refreshing instance network info cache due to event network-changed-787b8b7e-dea3-41b7-9349-4f87afe7b88d. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 792.782461] env[61985]: DEBUG oslo_concurrency.lockutils [req-aa2c0d3b-246c-4849-984d-72a11d08affa req-23aa2244-b230-4431-8486-2fa4a7e5fbb3 service nova] Acquiring lock "refresh_cache-f8bc8b71-0317-479d-b2f9-9471757f0774" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.782461] env[61985]: DEBUG oslo_concurrency.lockutils [req-aa2c0d3b-246c-4849-984d-72a11d08affa req-23aa2244-b230-4431-8486-2fa4a7e5fbb3 service nova] Acquired lock "refresh_cache-f8bc8b71-0317-479d-b2f9-9471757f0774" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.782461] env[61985]: DEBUG nova.network.neutron [req-aa2c0d3b-246c-4849-984d-72a11d08affa req-23aa2244-b230-4431-8486-2fa4a7e5fbb3 service nova] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Refreshing network info cache for port 787b8b7e-dea3-41b7-9349-4f87afe7b88d {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 792.902639] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935705, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.991596] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935706, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.950964} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.991712] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 3694e20c-ce37-4097-9991-8a06f38b2734/3694e20c-ce37-4097-9991-8a06f38b2734.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 792.991852] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.992120] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1efaef6a-19f0-4b49-b147-38fbcf869e00 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.998444] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 792.998444] env[61985]: value = "task-935710" [ 792.998444] env[61985]: _type = "Task" [ 792.998444] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.006465] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935710, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.033433] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5247df8e-ee55-5ecb-c83d-28e50276fe10, 'name': SearchDatastore_Task, 'duration_secs': 0.010255} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.034310] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35af582d-e8bd-427d-abc4-d73c8cd9c4c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.040300] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 793.040300] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c5e2b2-8269-c410-dbc5-e26cfcb62e6b" [ 793.040300] env[61985]: _type = "Task" [ 793.040300] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.049975] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c5e2b2-8269-c410-dbc5-e26cfcb62e6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.141247] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935709, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.289160] env[61985]: DEBUG nova.scheduler.client.report [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 74 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 793.289432] env[61985]: DEBUG nova.compute.provider_tree [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 74 to 75 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 793.289614] env[61985]: DEBUG nova.compute.provider_tree [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 793.307916] env[61985]: DEBUG nova.compute.manager [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 793.333537] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.333736] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.333897] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.334494] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.334494] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.334494] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.334670] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.334712] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.335605] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.335605] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.335605] env[61985]: DEBUG nova.virt.hardware [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.336883] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9757cde7-73fb-43ac-90bd-8464033492df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.344914] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9ae585-e410-4359-a827-602e70f9ed5d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.402771] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935705, 'name': Rename_Task, 'duration_secs': 1.136495} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.402935] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 793.403197] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42398faa-abad-4592-b02c-8bc226a4e400 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.409034] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for the task: (returnval){ [ 793.409034] env[61985]: value = "task-935711" [ 793.409034] env[61985]: _type = "Task" [ 793.409034] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.416341] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935711, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.510763] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935710, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107569} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.514237] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.515157] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03df8d99-7d63-449d-8d9a-94fd1f50cda6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.536973] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 3694e20c-ce37-4097-9991-8a06f38b2734/3694e20c-ce37-4097-9991-8a06f38b2734.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.537270] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3d3d940-cd02-488c-a077-2ea8defecd77 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.559387] env[61985]: DEBUG nova.compute.manager [req-1572d134-a028-409d-8b0c-48782a2efbb4 req-d67701db-f0de-409e-bebf-c4a10482c78c service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Received event network-vif-plugged-bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 793.559637] env[61985]: DEBUG oslo_concurrency.lockutils [req-1572d134-a028-409d-8b0c-48782a2efbb4 req-d67701db-f0de-409e-bebf-c4a10482c78c service nova] Acquiring lock "b659ff28-5101-4825-84ea-111351c81145-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.559901] env[61985]: DEBUG oslo_concurrency.lockutils [req-1572d134-a028-409d-8b0c-48782a2efbb4 req-d67701db-f0de-409e-bebf-c4a10482c78c service nova] Lock "b659ff28-5101-4825-84ea-111351c81145-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.560136] env[61985]: DEBUG oslo_concurrency.lockutils [req-1572d134-a028-409d-8b0c-48782a2efbb4 req-d67701db-f0de-409e-bebf-c4a10482c78c service nova] Lock "b659ff28-5101-4825-84ea-111351c81145-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.560362] env[61985]: DEBUG nova.compute.manager [req-1572d134-a028-409d-8b0c-48782a2efbb4 req-d67701db-f0de-409e-bebf-c4a10482c78c service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] No waiting events found dispatching network-vif-plugged-bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 793.560574] env[61985]: WARNING nova.compute.manager [req-1572d134-a028-409d-8b0c-48782a2efbb4 req-d67701db-f0de-409e-bebf-c4a10482c78c service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Received unexpected event network-vif-plugged-bf071457-a423-45ed-9a6e-22e52c290be8 for instance with vm_state building and task_state spawning. [ 793.562801] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 793.562801] env[61985]: value = "task-935712" [ 793.562801] env[61985]: _type = "Task" [ 793.562801] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.570904] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c5e2b2-8269-c410-dbc5-e26cfcb62e6b, 'name': SearchDatastore_Task, 'duration_secs': 0.01152} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.574528] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.574888] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 793.575176] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935712, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.575421] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32780e20-bc69-4ec1-a365-56e4fa7cf55c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.584988] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 793.584988] env[61985]: value = "task-935713" [ 793.584988] env[61985]: _type = "Task" [ 793.584988] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.592824] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935713, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.615118] env[61985]: DEBUG nova.network.neutron [req-aa2c0d3b-246c-4849-984d-72a11d08affa req-23aa2244-b230-4431-8486-2fa4a7e5fbb3 service nova] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Updated VIF entry in instance network info cache for port 787b8b7e-dea3-41b7-9349-4f87afe7b88d. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 793.615523] env[61985]: DEBUG nova.network.neutron [req-aa2c0d3b-246c-4849-984d-72a11d08affa req-23aa2244-b230-4431-8486-2fa4a7e5fbb3 service nova] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Updating instance_info_cache with network_info: [{"id": "787b8b7e-dea3-41b7-9349-4f87afe7b88d", "address": "fa:16:3e:2a:b2:dc", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap787b8b7e-de", "ovs_interfaceid": "787b8b7e-dea3-41b7-9349-4f87afe7b88d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.639081] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935709, 'name': CreateVM_Task, 'duration_secs': 0.924799} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.639343] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 793.639931] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.640064] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.640392] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.641190] env[61985]: DEBUG nova.network.neutron [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Successfully updated port: bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.645929] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-590bc8b0-c07d-4718-80ea-0807489a7e56 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.646963] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 793.646963] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c1156e-f682-b6d0-8e09-17c9c75ca89d" [ 793.646963] env[61985]: _type = "Task" [ 793.646963] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.657287] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c1156e-f682-b6d0-8e09-17c9c75ca89d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.794896] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.795678] env[61985]: DEBUG nova.compute.manager [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 793.798881] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.514s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.799095] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.799269] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 793.799558] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.562s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.801085] env[61985]: INFO nova.compute.claims [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.804319] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11826e2-0be7-4970-912f-91a7e3b9e11c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.812927] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb53d1bd-d44f-4fb3-bdba-669dd1f48c34 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.827557] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bfeed8a-61a5-4872-9a22-210d52c2043d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.835090] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5963b4c4-2b2f-409d-b79b-c4cb6c850357 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.864578] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181607MB free_disk=43GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 793.864753] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.918952] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935711, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.073574] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935712, 'name': ReconfigVM_Task, 'duration_secs': 0.49767} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.073893] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 3694e20c-ce37-4097-9991-8a06f38b2734/3694e20c-ce37-4097-9991-8a06f38b2734.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.074981] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f5abfa3-cf62-4161-bba6-e56fe1117765 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.081153] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 794.081153] env[61985]: value = "task-935714" [ 794.081153] env[61985]: _type = "Task" [ 794.081153] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.093130] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935714, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.099611] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935713, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.118665] env[61985]: DEBUG oslo_concurrency.lockutils [req-aa2c0d3b-246c-4849-984d-72a11d08affa req-23aa2244-b230-4431-8486-2fa4a7e5fbb3 service nova] Releasing lock "refresh_cache-f8bc8b71-0317-479d-b2f9-9471757f0774" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.145513] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.145751] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquired lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.145845] env[61985]: DEBUG nova.network.neutron [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 794.158876] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c1156e-f682-b6d0-8e09-17c9c75ca89d, 'name': SearchDatastore_Task, 'duration_secs': 0.02156} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.159863] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.160222] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 794.160465] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.160652] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.160857] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 794.161500] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa35df23-5576-44f8-b4c2-c704ec200203 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.175621] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 794.175882] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 794.176663] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eddfb930-2567-4dc7-b110-a529abf63651 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.184123] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 794.184123] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52baf5e5-a691-f8f2-5dcf-79199efce0d0" [ 794.184123] env[61985]: _type = "Task" [ 794.184123] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.193178] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52baf5e5-a691-f8f2-5dcf-79199efce0d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.305536] env[61985]: DEBUG nova.compute.utils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.307064] env[61985]: DEBUG nova.compute.manager [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 794.309558] env[61985]: DEBUG nova.network.neutron [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 794.367042] env[61985]: DEBUG nova.policy [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '00f4e0f073684affbdc22e676f440838', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '091a92048e3545b7a6b98af4e81d2f00', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 794.420648] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935711, 'name': PowerOnVM_Task} progress is 82%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.594996] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935714, 'name': Rename_Task, 'duration_secs': 0.192175} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.599335] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 794.599672] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935713, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.760601} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.599770] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6b9bd08-d567-4cc8-89f0-62d99e52c6b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.601868] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 794.601868] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 794.602062] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c68a7716-4a2e-4e67-a35f-3c3de86bc71d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.609856] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 794.609856] env[61985]: value = "task-935716" [ 794.609856] env[61985]: _type = "Task" [ 794.609856] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.611185] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 794.611185] env[61985]: value = "task-935715" [ 794.611185] env[61985]: _type = "Task" [ 794.611185] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.623983] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935715, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.627046] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935716, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.648216] env[61985]: DEBUG nova.network.neutron [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Successfully created port: 6ea4b2df-e4c4-48a4-9af2-fe4606759004 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.695658] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52baf5e5-a691-f8f2-5dcf-79199efce0d0, 'name': SearchDatastore_Task, 'duration_secs': 0.016393} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.703029] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e158d113-ab7e-47a8-bc30-217b4355c1f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.704223] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 794.704223] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c8edfb-4562-c5b9-1aff-6380770a3bc5" [ 794.704223] env[61985]: _type = "Task" [ 794.704223] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.708390] env[61985]: DEBUG nova.network.neutron [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 794.715368] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c8edfb-4562-c5b9-1aff-6380770a3bc5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.810150] env[61985]: DEBUG nova.compute.manager [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 794.898482] env[61985]: DEBUG nova.network.neutron [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Updating instance_info_cache with network_info: [{"id": "bf071457-a423-45ed-9a6e-22e52c290be8", "address": "fa:16:3e:8b:24:a0", "network": {"id": "a3e42a35-4a68-42be-bd75-5bf48aeab85b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1220560148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e8e226d5c9240928ccc35013189235c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf071457-a4", "ovs_interfaceid": "bf071457-a423-45ed-9a6e-22e52c290be8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.924263] env[61985]: DEBUG oslo_vmware.api [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935711, 'name': PowerOnVM_Task, 'duration_secs': 1.040182} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.924571] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 794.924748] env[61985]: INFO nova.compute.manager [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Took 9.88 seconds to spawn the instance on the hypervisor. [ 794.925604] env[61985]: DEBUG nova.compute.manager [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.925796] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b7a22d-10d2-4867-baa4-b40612f2114d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.126687] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935716, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074181} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.126950] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 795.131392] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd1ad6d9-63b6-4e50-8240-ad26483578b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.135461] env[61985]: DEBUG oslo_vmware.api [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935715, 'name': PowerOnVM_Task, 'duration_secs': 0.461939} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.135791] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 795.136057] env[61985]: INFO nova.compute.manager [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Took 7.53 seconds to spawn the instance on the hypervisor. [ 795.136278] env[61985]: DEBUG nova.compute.manager [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.137362] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a204db-637f-4893-8bb5-fe5940430201 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.159261] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 795.160325] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1fc793e-f88e-4ba1-8102-7545e846f65c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.187008] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 795.187008] env[61985]: value = "task-935717" [ 795.187008] env[61985]: _type = "Task" [ 795.187008] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.196389] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935717, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.213399] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c8edfb-4562-c5b9-1aff-6380770a3bc5, 'name': SearchDatastore_Task, 'duration_secs': 0.018584} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.213826] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.213909] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] f8bc8b71-0317-479d-b2f9-9471757f0774/f8bc8b71-0317-479d-b2f9-9471757f0774.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 795.214183] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0cb696d-a6e8-4f8a-ae48-6f135425473b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.220295] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 795.220295] env[61985]: value = "task-935718" [ 795.220295] env[61985]: _type = "Task" [ 795.220295] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.228789] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935718, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.255240] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970971f0-6227-4a33-8091-ee388a014aa3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.263204] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983a4a8b-4a10-498a-93f4-53124e131cd8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.293514] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57715ca-0b41-4a88-870b-97d4ff4b24a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.302654] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded27f08-0b3b-4368-a9a3-65be376f5976 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.317879] env[61985]: DEBUG nova.compute.provider_tree [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.407039] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Releasing lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.407039] env[61985]: DEBUG nova.compute.manager [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Instance network_info: |[{"id": "bf071457-a423-45ed-9a6e-22e52c290be8", "address": "fa:16:3e:8b:24:a0", "network": {"id": "a3e42a35-4a68-42be-bd75-5bf48aeab85b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1220560148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e8e226d5c9240928ccc35013189235c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf071457-a4", "ovs_interfaceid": "bf071457-a423-45ed-9a6e-22e52c290be8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 795.407379] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:24:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf071457-a423-45ed-9a6e-22e52c290be8', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 795.415112] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Creating folder: Project (8e8e226d5c9240928ccc35013189235c). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 795.415659] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-270eb039-4d20-4937-8f5b-6d8ad16e79b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.429209] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Created folder: Project (8e8e226d5c9240928ccc35013189235c) in parent group-v211285. [ 795.429209] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Creating folder: Instances. Parent ref: group-v211320. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 795.429209] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69373a7c-ff09-40e8-be77-69efbdd0a9fd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.439361] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Created folder: Instances in parent group-v211320. [ 795.440110] env[61985]: DEBUG oslo.service.loopingcall [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.440546] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b659ff28-5101-4825-84ea-111351c81145] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 795.446184] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a8149cab-773f-4a4a-8e4c-f0ceb9e6f74d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.473358] env[61985]: INFO nova.compute.manager [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Took 33.32 seconds to build instance. [ 795.478210] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 795.478210] env[61985]: value = "task-935721" [ 795.478210] env[61985]: _type = "Task" [ 795.478210] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.488569] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935721, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.693799] env[61985]: INFO nova.compute.manager [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Took 31.51 seconds to build instance. [ 795.699538] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935717, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.707101] env[61985]: DEBUG nova.compute.manager [req-c2facb55-dfa6-4f18-a7ac-d0464ef85e1d req-4d2c9f62-5a4e-4526-8813-4e6846317fe5 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Received event network-changed-bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 795.707817] env[61985]: DEBUG nova.compute.manager [req-c2facb55-dfa6-4f18-a7ac-d0464ef85e1d req-4d2c9f62-5a4e-4526-8813-4e6846317fe5 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Refreshing instance network info cache due to event network-changed-bf071457-a423-45ed-9a6e-22e52c290be8. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 795.707817] env[61985]: DEBUG oslo_concurrency.lockutils [req-c2facb55-dfa6-4f18-a7ac-d0464ef85e1d req-4d2c9f62-5a4e-4526-8813-4e6846317fe5 service nova] Acquiring lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.707817] env[61985]: DEBUG oslo_concurrency.lockutils [req-c2facb55-dfa6-4f18-a7ac-d0464ef85e1d req-4d2c9f62-5a4e-4526-8813-4e6846317fe5 service nova] Acquired lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.708227] env[61985]: DEBUG nova.network.neutron [req-c2facb55-dfa6-4f18-a7ac-d0464ef85e1d req-4d2c9f62-5a4e-4526-8813-4e6846317fe5 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Refreshing network info cache for port bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 795.730390] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935718, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459885} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.730645] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] f8bc8b71-0317-479d-b2f9-9471757f0774/f8bc8b71-0317-479d-b2f9-9471757f0774.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 795.730890] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.731159] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c269beaa-1f45-4d70-ab01-9d6b6b6979cd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.738193] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 795.738193] env[61985]: value = "task-935722" [ 795.738193] env[61985]: _type = "Task" [ 795.738193] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.749630] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935722, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.821103] env[61985]: DEBUG nova.scheduler.client.report [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.829100] env[61985]: DEBUG nova.compute.manager [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 795.866772] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.867155] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.867424] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.867708] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.867919] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.868134] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.868406] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.868628] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.868896] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.869135] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.869377] env[61985]: DEBUG nova.virt.hardware [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.870309] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668a7ede-61b5-4c7f-8412-4e4cf1e9bda9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.882030] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eca0f4e-4f07-46cd-a782-a71c3b920ce4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.977240] env[61985]: DEBUG oslo_concurrency.lockutils [None req-764feb47-f908-4772-8197-86e81f1bfecc tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.164s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.990905] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935721, 'name': CreateVM_Task, 'duration_secs': 0.4501} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.990905] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b659ff28-5101-4825-84ea-111351c81145] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 795.990905] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.991438] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.991868] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 795.992729] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3dd55d08-62aa-4826-ae5a-8ab137316183 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.998181] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 795.998181] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f04cad-fcc6-ebb1-2cb5-f75897644a70" [ 795.998181] env[61985]: _type = "Task" [ 795.998181] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.007392] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f04cad-fcc6-ebb1-2cb5-f75897644a70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.196431] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b422d591-e070-494c-81e3-88d661654f13 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "3694e20c-ce37-4097-9991-8a06f38b2734" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.421s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.205957] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935717, 'name': ReconfigVM_Task, 'duration_secs': 0.533229} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.206417] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 796.207424] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e951272-3307-45a3-abc3-421ab96d5cd8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.210561] env[61985]: DEBUG nova.network.neutron [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Successfully updated port: 6ea4b2df-e4c4-48a4-9af2-fe4606759004 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.222118] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 796.222118] env[61985]: value = "task-935723" [ 796.222118] env[61985]: _type = "Task" [ 796.222118] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.231338] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935723, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.247823] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935722, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070545} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.248286] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.251131] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26271f7-80fc-4234-b5b7-64050499a12a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.274515] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] f8bc8b71-0317-479d-b2f9-9471757f0774/f8bc8b71-0317-479d-b2f9-9471757f0774.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.275794] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d586d75c-c728-4dea-8fa1-94b16d6be70d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.298230] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 796.298230] env[61985]: value = "task-935724" [ 796.298230] env[61985]: _type = "Task" [ 796.298230] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.307235] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935724, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.335957] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.336838] env[61985]: DEBUG nova.compute.manager [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 796.339401] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.843s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.340935] env[61985]: INFO nova.compute.claims [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.447680] env[61985]: DEBUG nova.network.neutron [req-c2facb55-dfa6-4f18-a7ac-d0464ef85e1d req-4d2c9f62-5a4e-4526-8813-4e6846317fe5 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Updated VIF entry in instance network info cache for port bf071457-a423-45ed-9a6e-22e52c290be8. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 796.448142] env[61985]: DEBUG nova.network.neutron [req-c2facb55-dfa6-4f18-a7ac-d0464ef85e1d req-4d2c9f62-5a4e-4526-8813-4e6846317fe5 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Updating instance_info_cache with network_info: [{"id": "bf071457-a423-45ed-9a6e-22e52c290be8", "address": "fa:16:3e:8b:24:a0", "network": {"id": "a3e42a35-4a68-42be-bd75-5bf48aeab85b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1220560148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e8e226d5c9240928ccc35013189235c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf071457-a4", "ovs_interfaceid": "bf071457-a423-45ed-9a6e-22e52c290be8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.479307] env[61985]: DEBUG nova.compute.manager [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.509544] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f04cad-fcc6-ebb1-2cb5-f75897644a70, 'name': SearchDatastore_Task, 'duration_secs': 0.036812} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.509853] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.510100] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.510335] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.510479] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.510659] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.510917] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a50bac72-56b4-4305-b7b4-0fba37dc6451 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.520509] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.520741] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 796.521736] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-874afcc9-89e2-4497-a1bc-7fc8ae1406c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.529422] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 796.529422] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a1b61b-1d38-bc3d-6f0a-fc3282491007" [ 796.529422] env[61985]: _type = "Task" [ 796.529422] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.537286] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a1b61b-1d38-bc3d-6f0a-fc3282491007, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.630130] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquiring lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.630439] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.630717] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquiring lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.630923] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.631112] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.633468] env[61985]: INFO nova.compute.manager [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Terminating instance [ 796.635447] env[61985]: DEBUG nova.compute.manager [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 796.635803] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 796.636646] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6220bf-583c-48e5-817a-2c39f60864d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.644812] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 796.644971] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd50b549-8a2b-40c4-8232-28acfa6d5a6c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.651743] env[61985]: DEBUG oslo_vmware.api [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for the task: (returnval){ [ 796.651743] env[61985]: value = "task-935725" [ 796.651743] env[61985]: _type = "Task" [ 796.651743] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.659581] env[61985]: DEBUG oslo_vmware.api [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935725, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.702280] env[61985]: DEBUG nova.compute.manager [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.717926] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.718940] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.718940] env[61985]: DEBUG nova.network.neutron [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.731673] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935723, 'name': Rename_Task, 'duration_secs': 0.388654} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.732530] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 796.732847] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28994038-1c0f-4da2-bfb6-de868a8c75d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.739771] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 796.739771] env[61985]: value = "task-935726" [ 796.739771] env[61985]: _type = "Task" [ 796.739771] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.748929] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935726, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.809046] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935724, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.845557] env[61985]: DEBUG nova.compute.utils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.849407] env[61985]: DEBUG nova.compute.manager [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 796.849609] env[61985]: DEBUG nova.network.neutron [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 796.891935] env[61985]: DEBUG nova.policy [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '30ab573be3914403a9d28db9a9897d12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd536218a37cb4de387cad646c8533695', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.951806] env[61985]: DEBUG oslo_concurrency.lockutils [req-c2facb55-dfa6-4f18-a7ac-d0464ef85e1d req-4d2c9f62-5a4e-4526-8813-4e6846317fe5 service nova] Releasing lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.007915] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.040420] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a1b61b-1d38-bc3d-6f0a-fc3282491007, 'name': SearchDatastore_Task, 'duration_secs': 0.009231} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.042384] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f35fd092-5694-4b77-acbf-86849cff0b3b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.047391] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 797.047391] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a8de22-4ed4-240b-af86-8de3b09bab82" [ 797.047391] env[61985]: _type = "Task" [ 797.047391] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.056501] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a8de22-4ed4-240b-af86-8de3b09bab82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.165081] env[61985]: DEBUG oslo_vmware.api [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935725, 'name': PowerOffVM_Task, 'duration_secs': 0.391649} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.165421] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 797.165650] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 797.165938] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-314e57c9-6769-4f7c-8a4e-6bd946f17414 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.168850] env[61985]: DEBUG nova.network.neutron [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Successfully created port: 9beae88f-5110-4296-bf01-5b1c423e9017 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.227247] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.245440] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 797.245703] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 797.245890] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Deleting the datastore file [datastore1] 3815af53-ba3b-4c60-a5fd-2d94102cdd42 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.246568] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc9d7817-90c8-4b8f-b7b8-b4089b977126 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.252642] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935726, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.258393] env[61985]: DEBUG oslo_vmware.api [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for the task: (returnval){ [ 797.258393] env[61985]: value = "task-935728" [ 797.258393] env[61985]: _type = "Task" [ 797.258393] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.269655] env[61985]: DEBUG oslo_vmware.api [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935728, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.276912] env[61985]: DEBUG nova.network.neutron [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.309603] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935724, 'name': ReconfigVM_Task, 'duration_secs': 1.01157} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.310092] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Reconfigured VM instance instance-0000002b to attach disk [datastore1] f8bc8b71-0317-479d-b2f9-9471757f0774/f8bc8b71-0317-479d-b2f9-9471757f0774.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.310568] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-47fcca24-14d6-4697-86fd-9612773274d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.317624] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 797.317624] env[61985]: value = "task-935729" [ 797.317624] env[61985]: _type = "Task" [ 797.317624] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.327230] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935729, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.350088] env[61985]: DEBUG nova.compute.manager [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 797.561172] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a8de22-4ed4-240b-af86-8de3b09bab82, 'name': SearchDatastore_Task, 'duration_secs': 0.010448} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.564110] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.567667] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] b659ff28-5101-4825-84ea-111351c81145/b659ff28-5101-4825-84ea-111351c81145.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 797.568608] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0332e2be-7d29-4373-9bb3-254adc94a609 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.572896] env[61985]: DEBUG nova.network.neutron [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Updating instance_info_cache with network_info: [{"id": "6ea4b2df-e4c4-48a4-9af2-fe4606759004", "address": "fa:16:3e:b3:50:b6", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ea4b2df-e4", "ovs_interfaceid": "6ea4b2df-e4c4-48a4-9af2-fe4606759004", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.579640] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 797.579640] env[61985]: value = "task-935730" [ 797.579640] env[61985]: _type = "Task" [ 797.579640] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.593919] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935730, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.757655] env[61985]: DEBUG oslo_vmware.api [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935726, 'name': PowerOnVM_Task, 'duration_secs': 0.885928} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.758098] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 797.758429] env[61985]: DEBUG nova.compute.manager [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 797.759276] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d15b3f8-cf34-435c-af1b-f8612101eef4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.776469] env[61985]: DEBUG oslo_vmware.api [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Task: {'id': task-935728, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134431} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.777129] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.777353] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 797.777559] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 797.777749] env[61985]: INFO nova.compute.manager [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Took 1.14 seconds to destroy the instance on the hypervisor. [ 797.777992] env[61985]: DEBUG oslo.service.loopingcall [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.778236] env[61985]: DEBUG nova.compute.manager [-] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.778333] env[61985]: DEBUG nova.network.neutron [-] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 797.812367] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d6273c-e88f-4d92-900c-a74ba3ab592d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.824450] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db7ac33-1c7f-4301-ae20-12ed6c47c1cd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.833426] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935729, 'name': Rename_Task, 'duration_secs': 0.149812} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.860914] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 797.868619] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb278acc-49a2-4217-944d-2b66ae18a76c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.871210] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133b623f-99a7-4006-ae55-b2dd56483df7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.881348] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-860abd1a-9397-4ae0-9ff2-e111bb6dbeab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.885810] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 797.885810] env[61985]: value = "task-935731" [ 797.885810] env[61985]: _type = "Task" [ 797.885810] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.901473] env[61985]: DEBUG nova.compute.provider_tree [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.909204] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935731, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.080029] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.080029] env[61985]: DEBUG nova.compute.manager [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Instance network_info: |[{"id": "6ea4b2df-e4c4-48a4-9af2-fe4606759004", "address": "fa:16:3e:b3:50:b6", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ea4b2df-e4", "ovs_interfaceid": "6ea4b2df-e4c4-48a4-9af2-fe4606759004", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 798.080651] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:50:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cd38fdec-d092-4a84-ab41-685f6dbb4f29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ea4b2df-e4c4-48a4-9af2-fe4606759004', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.086733] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Creating folder: Project (091a92048e3545b7a6b98af4e81d2f00). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 798.087273] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-02b8e221-95fd-4656-9044-446c0d739e6b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.098365] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935730, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476557} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.098876] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] b659ff28-5101-4825-84ea-111351c81145/b659ff28-5101-4825-84ea-111351c81145.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 798.099283] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.099661] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d19d20b-98c0-4d84-924c-6540dd6a521a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.103023] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Created folder: Project (091a92048e3545b7a6b98af4e81d2f00) in parent group-v211285. [ 798.103393] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Creating folder: Instances. Parent ref: group-v211323. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 798.104088] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-000cfa54-265f-47fc-9968-14b251537bd8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.107841] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 798.107841] env[61985]: value = "task-935733" [ 798.107841] env[61985]: _type = "Task" [ 798.107841] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.116019] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Created folder: Instances in parent group-v211323. [ 798.116019] env[61985]: DEBUG oslo.service.loopingcall [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.117711] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 798.118421] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935733, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.118742] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c530420-a37f-4c89-8ac4-93d6277e38fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.140195] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.140195] env[61985]: value = "task-935735" [ 798.140195] env[61985]: _type = "Task" [ 798.140195] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.150784] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935735, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.224945] env[61985]: DEBUG nova.compute.manager [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Received event network-vif-plugged-6ea4b2df-e4c4-48a4-9af2-fe4606759004 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 798.224945] env[61985]: DEBUG oslo_concurrency.lockutils [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] Acquiring lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.227216] env[61985]: DEBUG oslo_concurrency.lockutils [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] Lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.227442] env[61985]: DEBUG oslo_concurrency.lockutils [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] Lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.227696] env[61985]: DEBUG nova.compute.manager [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] No waiting events found dispatching network-vif-plugged-6ea4b2df-e4c4-48a4-9af2-fe4606759004 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 798.227911] env[61985]: WARNING nova.compute.manager [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Received unexpected event network-vif-plugged-6ea4b2df-e4c4-48a4-9af2-fe4606759004 for instance with vm_state building and task_state spawning. [ 798.228248] env[61985]: DEBUG nova.compute.manager [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Received event network-changed-6ea4b2df-e4c4-48a4-9af2-fe4606759004 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 798.228337] env[61985]: DEBUG nova.compute.manager [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Refreshing instance network info cache due to event network-changed-6ea4b2df-e4c4-48a4-9af2-fe4606759004. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 798.228520] env[61985]: DEBUG oslo_concurrency.lockutils [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] Acquiring lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.228677] env[61985]: DEBUG oslo_concurrency.lockutils [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] Acquired lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.228876] env[61985]: DEBUG nova.network.neutron [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Refreshing network info cache for port 6ea4b2df-e4c4-48a4-9af2-fe4606759004 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.293301] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.370414] env[61985]: DEBUG nova.compute.manager [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 798.399916] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935731, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.403487] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.403810] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.404026] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.404422] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.404665] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.404872] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.405421] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.405665] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.405887] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.406103] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.406512] env[61985]: DEBUG nova.virt.hardware [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.407698] env[61985]: DEBUG nova.scheduler.client.report [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.412218] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca49ea68-8b9c-421a-81ae-11b4d6e08a74 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.420557] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f8d901-48bf-4950-9665-77caa158bd28 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.535193] env[61985]: DEBUG nova.compute.manager [req-532f0cf4-0c1d-4f2e-8ee0-fcfa9e67a4dd req-a6d7f96f-eaf1-409f-8101-494d7a3e4d73 service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Received event network-vif-deleted-cdd07618-1786-48b0-9f82-2fe473047b27 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 798.535193] env[61985]: INFO nova.compute.manager [req-532f0cf4-0c1d-4f2e-8ee0-fcfa9e67a4dd req-a6d7f96f-eaf1-409f-8101-494d7a3e4d73 service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Neutron deleted interface cdd07618-1786-48b0-9f82-2fe473047b27; detaching it from the instance and deleting it from the info cache [ 798.535193] env[61985]: DEBUG nova.network.neutron [req-532f0cf4-0c1d-4f2e-8ee0-fcfa9e67a4dd req-a6d7f96f-eaf1-409f-8101-494d7a3e4d73 service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.568554] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "c64056d8-0c22-48db-bf4a-a60f468bdad0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.568554] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.620651] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935733, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080005} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.621035] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 798.621856] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0520b146-90fd-4c36-8646-ae19afad44c8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.647020] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] b659ff28-5101-4825-84ea-111351c81145/b659ff28-5101-4825-84ea-111351c81145.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.647321] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53cb86dc-b95c-4c34-b2f6-70e36d39e670 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.670107] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935735, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.671475] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 798.671475] env[61985]: value = "task-935736" [ 798.671475] env[61985]: _type = "Task" [ 798.671475] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.679365] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935736, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.735818] env[61985]: DEBUG nova.network.neutron [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Successfully updated port: 9beae88f-5110-4296-bf01-5b1c423e9017 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.761591] env[61985]: DEBUG nova.network.neutron [-] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.896816] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935731, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.913775] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.914513] env[61985]: DEBUG nova.compute.manager [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 798.917649] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.783s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.919035] env[61985]: INFO nova.compute.claims [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.040608] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bae12a54-c275-48da-b6de-7af1c847ebb0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.053069] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c071873-e11f-42b8-962b-7d7a1272a124 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.077568] env[61985]: DEBUG nova.compute.manager [req-532f0cf4-0c1d-4f2e-8ee0-fcfa9e67a4dd req-a6d7f96f-eaf1-409f-8101-494d7a3e4d73 service nova] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Detach interface failed, port_id=cdd07618-1786-48b0-9f82-2fe473047b27, reason: Instance 3815af53-ba3b-4c60-a5fd-2d94102cdd42 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 799.156535] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935735, 'name': CreateVM_Task, 'duration_secs': 0.754577} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.156807] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 799.157576] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.157663] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.158031] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.158265] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56acfd11-e286-44f1-b122-72fe11fbc701 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.162580] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 799.162580] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5287b39e-b0db-7e9f-2fb7-1993b452cf81" [ 799.162580] env[61985]: _type = "Task" [ 799.162580] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.170093] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5287b39e-b0db-7e9f-2fb7-1993b452cf81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.179559] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935736, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.233253] env[61985]: DEBUG nova.network.neutron [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Updated VIF entry in instance network info cache for port 6ea4b2df-e4c4-48a4-9af2-fe4606759004. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 799.233624] env[61985]: DEBUG nova.network.neutron [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Updating instance_info_cache with network_info: [{"id": "6ea4b2df-e4c4-48a4-9af2-fe4606759004", "address": "fa:16:3e:b3:50:b6", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ea4b2df-e4", "ovs_interfaceid": "6ea4b2df-e4c4-48a4-9af2-fe4606759004", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.243049] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquiring lock "refresh_cache-4aca0ecb-4ae6-4400-accd-d71782b2806d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.243049] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquired lock "refresh_cache-4aca0ecb-4ae6-4400-accd-d71782b2806d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.243186] env[61985]: DEBUG nova.network.neutron [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 799.264295] env[61985]: INFO nova.compute.manager [-] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Took 1.49 seconds to deallocate network for instance. [ 799.396318] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935731, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.423760] env[61985]: DEBUG nova.compute.utils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.427954] env[61985]: DEBUG nova.compute.manager [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 799.428186] env[61985]: DEBUG nova.network.neutron [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 799.487940] env[61985]: DEBUG nova.policy [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '00f4e0f073684affbdc22e676f440838', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '091a92048e3545b7a6b98af4e81d2f00', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 799.673106] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5287b39e-b0db-7e9f-2fb7-1993b452cf81, 'name': SearchDatastore_Task, 'duration_secs': 0.010832} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.676205] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.676454] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.676753] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.676916] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.677116] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.677368] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-72f52005-13cf-4b10-b6d6-2f025d5591a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.687853] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935736, 'name': ReconfigVM_Task, 'duration_secs': 0.744923} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.688155] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Reconfigured VM instance instance-0000002c to attach disk [datastore2] b659ff28-5101-4825-84ea-111351c81145/b659ff28-5101-4825-84ea-111351c81145.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.688989] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3119ffa2-ad91-4d0b-9862-e8906da407bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.692172] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.692353] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 799.693418] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf90d21c-1ae1-4488-8f69-b2bdef84458c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.698240] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 799.698240] env[61985]: value = "task-935737" [ 799.698240] env[61985]: _type = "Task" [ 799.698240] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.702854] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 799.702854] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525f010a-a1ad-a416-ef98-766167cde62e" [ 799.702854] env[61985]: _type = "Task" [ 799.702854] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.709199] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935737, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.713311] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525f010a-a1ad-a416-ef98-766167cde62e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.736122] env[61985]: DEBUG oslo_concurrency.lockutils [req-840e88fb-d619-4bdb-a8fb-a4295d3ce8bc req-58d05cfc-72a0-4e08-8f54-77355a20c829 service nova] Releasing lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.771049] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.784029] env[61985]: DEBUG nova.network.neutron [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.872513] env[61985]: INFO nova.compute.manager [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Rebuilding instance [ 799.903488] env[61985]: DEBUG oslo_vmware.api [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935731, 'name': PowerOnVM_Task, 'duration_secs': 1.599988} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.907480] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 799.907732] env[61985]: INFO nova.compute.manager [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Took 9.79 seconds to spawn the instance on the hypervisor. [ 799.907938] env[61985]: DEBUG nova.compute.manager [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.909566] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3029f4df-108e-4707-bdd6-26112fa49ab6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.928879] env[61985]: DEBUG nova.compute.manager [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 799.936601] env[61985]: DEBUG nova.compute.manager [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.936808] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6855a8-586b-4eeb-8340-303dc4717fa2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.942290] env[61985]: DEBUG nova.network.neutron [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Updating instance_info_cache with network_info: [{"id": "9beae88f-5110-4296-bf01-5b1c423e9017", "address": "fa:16:3e:8e:7f:19", "network": {"id": "c0bbb2c0-0de9-4369-9d51-bd03840c7215", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-61622794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d536218a37cb4de387cad646c8533695", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9beae88f-51", "ovs_interfaceid": "9beae88f-5110-4296-bf01-5b1c423e9017", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.950296] env[61985]: DEBUG nova.network.neutron [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Successfully created port: c529f7f6-f908-4a69-944f-b979f622f99f {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.213888] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935737, 'name': Rename_Task, 'duration_secs': 0.299108} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.219871] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 800.220200] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525f010a-a1ad-a416-ef98-766167cde62e, 'name': SearchDatastore_Task, 'duration_secs': 0.019003} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.220591] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d54ade1d-47ab-42c9-a78a-141b6862e740 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.222631] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4bc02fc-3e4e-492b-a293-77be5d4bf25d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.229021] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 800.229021] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5210bbe9-806f-a578-00b4-c1d82695f088" [ 800.229021] env[61985]: _type = "Task" [ 800.229021] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.229722] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 800.229722] env[61985]: value = "task-935738" [ 800.229722] env[61985]: _type = "Task" [ 800.229722] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.243223] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5210bbe9-806f-a578-00b4-c1d82695f088, 'name': SearchDatastore_Task, 'duration_secs': 0.010734} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.246621] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.246752] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 761d5954-a5ca-4459-a1d6-bfc59b284bf4/761d5954-a5ca-4459-a1d6-bfc59b284bf4.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 800.247034] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935738, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.247251] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da9ab6e9-2050-4d2c-a498-76f8cfa3565f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.256294] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 800.256294] env[61985]: value = "task-935739" [ 800.256294] env[61985]: _type = "Task" [ 800.256294] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.266735] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935739, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.380195] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ebe504-8074-4c28-a761-e28daf13062a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.392614] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88745984-7da8-4ab3-ab96-64bbb5ab3900 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.431916] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d73c0f-2a48-4540-9ad0-b015dc9247d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.435976] env[61985]: DEBUG nova.compute.manager [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Received event network-vif-plugged-9beae88f-5110-4296-bf01-5b1c423e9017 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 800.436367] env[61985]: DEBUG oslo_concurrency.lockutils [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] Acquiring lock "4aca0ecb-4ae6-4400-accd-d71782b2806d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.436657] env[61985]: DEBUG oslo_concurrency.lockutils [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] Lock "4aca0ecb-4ae6-4400-accd-d71782b2806d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.436872] env[61985]: DEBUG oslo_concurrency.lockutils [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] Lock "4aca0ecb-4ae6-4400-accd-d71782b2806d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.437094] env[61985]: DEBUG nova.compute.manager [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] No waiting events found dispatching network-vif-plugged-9beae88f-5110-4296-bf01-5b1c423e9017 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 800.437297] env[61985]: WARNING nova.compute.manager [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Received unexpected event network-vif-plugged-9beae88f-5110-4296-bf01-5b1c423e9017 for instance with vm_state building and task_state spawning. [ 800.437478] env[61985]: DEBUG nova.compute.manager [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Received event network-changed-9beae88f-5110-4296-bf01-5b1c423e9017 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 800.437699] env[61985]: DEBUG nova.compute.manager [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Refreshing instance network info cache due to event network-changed-9beae88f-5110-4296-bf01-5b1c423e9017. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 800.437908] env[61985]: DEBUG oslo_concurrency.lockutils [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] Acquiring lock "refresh_cache-4aca0ecb-4ae6-4400-accd-d71782b2806d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.438438] env[61985]: INFO nova.compute.manager [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Took 34.22 seconds to build instance. [ 800.444958] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Releasing lock "refresh_cache-4aca0ecb-4ae6-4400-accd-d71782b2806d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.445299] env[61985]: DEBUG nova.compute.manager [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Instance network_info: |[{"id": "9beae88f-5110-4296-bf01-5b1c423e9017", "address": "fa:16:3e:8e:7f:19", "network": {"id": "c0bbb2c0-0de9-4369-9d51-bd03840c7215", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-61622794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d536218a37cb4de387cad646c8533695", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9beae88f-51", "ovs_interfaceid": "9beae88f-5110-4296-bf01-5b1c423e9017", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.445608] env[61985]: DEBUG oslo_concurrency.lockutils [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] Acquired lock "refresh_cache-4aca0ecb-4ae6-4400-accd-d71782b2806d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.445831] env[61985]: DEBUG nova.network.neutron [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Refreshing network info cache for port 9beae88f-5110-4296-bf01-5b1c423e9017 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.446967] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:7f:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9beae88f-5110-4296-bf01-5b1c423e9017', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.455335] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Creating folder: Project (d536218a37cb4de387cad646c8533695). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 800.461442] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 800.461442] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9daf07d5-afa9-4ac1-9567-bedbb183d251 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.462498] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69e3572d-abf4-4c16-91ae-d8d18836c9f7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.465345] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f161ec5b-5b3b-4c3a-a3c4-1ff70b0179d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.483474] env[61985]: DEBUG nova.compute.provider_tree [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.487789] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for the task: (returnval){ [ 800.487789] env[61985]: value = "task-935741" [ 800.487789] env[61985]: _type = "Task" [ 800.487789] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.489163] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Created folder: Project (d536218a37cb4de387cad646c8533695) in parent group-v211285. [ 800.489357] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Creating folder: Instances. Parent ref: group-v211326. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 800.489908] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a48afb7b-16bb-4cdc-9b64-b47df57f17f5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.502264] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935741, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.503948] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Created folder: Instances in parent group-v211326. [ 800.504219] env[61985]: DEBUG oslo.service.loopingcall [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.504407] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 800.504617] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0d8b4c6-0c98-4d42-8b66-70f5e712f50c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.525096] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.525096] env[61985]: value = "task-935743" [ 800.525096] env[61985]: _type = "Task" [ 800.525096] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.534219] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935743, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.743699] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935738, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.766400] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935739, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.945684] env[61985]: DEBUG oslo_concurrency.lockutils [None req-566815a2-980f-49fa-837c-5ca536ecb454 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "f8bc8b71-0317-479d-b2f9-9471757f0774" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.207s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.948778] env[61985]: DEBUG nova.compute.manager [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 800.974235] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.974235] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.974235] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.974404] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.974556] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.974702] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.974932] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.975741] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.975843] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.975941] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.976148] env[61985]: DEBUG nova.virt.hardware [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.981030] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc5ad0d-429c-4f33-9960-3b0eede902c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.986309] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51887971-7712-4473-9ca1-b8909c0c6eec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.992700] env[61985]: DEBUG nova.scheduler.client.report [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.013158] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935741, 'name': PowerOffVM_Task, 'duration_secs': 0.202751} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.013488] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 801.013619] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 801.014504] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba90578-29fb-4d0b-be6e-648edd8ecccc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.022473] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 801.022713] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfc40265-8319-4384-b001-f0fecc6dbc4a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.034608] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935743, 'name': CreateVM_Task, 'duration_secs': 0.415864} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.034608] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 801.037356] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.038250] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.038250] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.040382] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83dc67d1-5bba-4129-aed3-e91c9b85b987 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.045845] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for the task: (returnval){ [ 801.045845] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b6c69f-0743-9b5c-8f9f-9813c8e1e446" [ 801.045845] env[61985]: _type = "Task" [ 801.045845] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.053085] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 801.053085] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 801.053204] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Deleting the datastore file [datastore2] 794cff68-6e26-4607-96f7-eaeb41182551 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 801.056072] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f86c8a36-919a-49e0-badd-7d7c5d0d8ed0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.057789] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b6c69f-0743-9b5c-8f9f-9813c8e1e446, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.064740] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for the task: (returnval){ [ 801.064740] env[61985]: value = "task-935745" [ 801.064740] env[61985]: _type = "Task" [ 801.064740] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.072372] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935745, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.184199] env[61985]: DEBUG nova.network.neutron [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Updated VIF entry in instance network info cache for port 9beae88f-5110-4296-bf01-5b1c423e9017. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.184562] env[61985]: DEBUG nova.network.neutron [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Updating instance_info_cache with network_info: [{"id": "9beae88f-5110-4296-bf01-5b1c423e9017", "address": "fa:16:3e:8e:7f:19", "network": {"id": "c0bbb2c0-0de9-4369-9d51-bd03840c7215", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-61622794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d536218a37cb4de387cad646c8533695", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9beae88f-51", "ovs_interfaceid": "9beae88f-5110-4296-bf01-5b1c423e9017", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.242177] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935738, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.267815] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935739, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528012} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.268206] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 761d5954-a5ca-4459-a1d6-bfc59b284bf4/761d5954-a5ca-4459-a1d6-bfc59b284bf4.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 801.268472] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.268784] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3e358d5-f654-45b7-9f46-60d8e647c524 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.277183] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 801.277183] env[61985]: value = "task-935746" [ 801.277183] env[61985]: _type = "Task" [ 801.277183] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.286329] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935746, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.453152] env[61985]: DEBUG nova.compute.manager [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 801.501240] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.584s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.501769] env[61985]: DEBUG nova.compute.manager [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 801.504650] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.370s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.506156] env[61985]: INFO nova.compute.claims [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.562084] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b6c69f-0743-9b5c-8f9f-9813c8e1e446, 'name': SearchDatastore_Task, 'duration_secs': 0.011421} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.562692] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.563431] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.563818] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.564314] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.564618] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.565097] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e11f00fc-6f25-4c7f-b3f5-0e7ace3538bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.578870] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935745, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153819} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.580540] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 801.580797] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 801.581062] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 801.583560] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.583774] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 801.585128] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a054aad-26ef-4514-b429-e3df34ae6eda {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.591307] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for the task: (returnval){ [ 801.591307] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521417ba-ccb5-2baf-bda4-1869a0d20abe" [ 801.591307] env[61985]: _type = "Task" [ 801.591307] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.600667] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521417ba-ccb5-2baf-bda4-1869a0d20abe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.687037] env[61985]: DEBUG oslo_concurrency.lockutils [req-588fabf4-8426-49e4-9e43-c2ada7e33773 req-6cfea8c9-78cc-4e42-93aa-a082550e1689 service nova] Releasing lock "refresh_cache-4aca0ecb-4ae6-4400-accd-d71782b2806d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.745519] env[61985]: DEBUG oslo_vmware.api [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935738, 'name': PowerOnVM_Task, 'duration_secs': 1.168927} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.745820] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 801.746646] env[61985]: INFO nova.compute.manager [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Took 8.44 seconds to spawn the instance on the hypervisor. [ 801.746646] env[61985]: DEBUG nova.compute.manager [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 801.747035] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f800be01-2602-422c-bd46-57a003a235bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.788168] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935746, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.195595} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.788403] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.789161] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3251db55-8c42-4607-9c76-888a2a6f6ffd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.815017] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 761d5954-a5ca-4459-a1d6-bfc59b284bf4/761d5954-a5ca-4459-a1d6-bfc59b284bf4.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.816191] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04cefe47-f42c-4cb8-9e4b-387ab72c723b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.838463] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 801.838463] env[61985]: value = "task-935747" [ 801.838463] env[61985]: _type = "Task" [ 801.838463] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.846873] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935747, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.952668] env[61985]: INFO nova.compute.manager [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Rebuilding instance [ 801.985738] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.999745] env[61985]: DEBUG nova.compute.manager [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 802.000638] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768feb57-e10c-46f3-bc74-b65e7515a07c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.009907] env[61985]: DEBUG nova.compute.utils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.016893] env[61985]: DEBUG nova.compute.manager [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.017084] env[61985]: DEBUG nova.network.neutron [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 802.029724] env[61985]: DEBUG nova.network.neutron [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Successfully updated port: c529f7f6-f908-4a69-944f-b979f622f99f {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 802.047850] env[61985]: DEBUG nova.compute.manager [req-38151871-108b-47d8-88f2-b6cbca40c60e req-5af51bb1-de1d-4cb5-bc6f-0b38c5a95893 service nova] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Received event network-vif-plugged-c529f7f6-f908-4a69-944f-b979f622f99f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 802.048090] env[61985]: DEBUG oslo_concurrency.lockutils [req-38151871-108b-47d8-88f2-b6cbca40c60e req-5af51bb1-de1d-4cb5-bc6f-0b38c5a95893 service nova] Acquiring lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.048307] env[61985]: DEBUG oslo_concurrency.lockutils [req-38151871-108b-47d8-88f2-b6cbca40c60e req-5af51bb1-de1d-4cb5-bc6f-0b38c5a95893 service nova] Lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.048549] env[61985]: DEBUG oslo_concurrency.lockutils [req-38151871-108b-47d8-88f2-b6cbca40c60e req-5af51bb1-de1d-4cb5-bc6f-0b38c5a95893 service nova] Lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.048747] env[61985]: DEBUG nova.compute.manager [req-38151871-108b-47d8-88f2-b6cbca40c60e req-5af51bb1-de1d-4cb5-bc6f-0b38c5a95893 service nova] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] No waiting events found dispatching network-vif-plugged-c529f7f6-f908-4a69-944f-b979f622f99f {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 802.048986] env[61985]: WARNING nova.compute.manager [req-38151871-108b-47d8-88f2-b6cbca40c60e req-5af51bb1-de1d-4cb5-bc6f-0b38c5a95893 service nova] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Received unexpected event network-vif-plugged-c529f7f6-f908-4a69-944f-b979f622f99f for instance with vm_state building and task_state spawning. [ 802.065047] env[61985]: DEBUG nova.policy [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faa19f385cd0426d8565e31e286800d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5625013599b44418bd56eb604e14be58', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 802.103538] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521417ba-ccb5-2baf-bda4-1869a0d20abe, 'name': SearchDatastore_Task, 'duration_secs': 0.010352} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.104438] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-380c5d2e-5b0d-491c-ab1b-91034d55135f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.110365] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for the task: (returnval){ [ 802.110365] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b296a9-a9a8-b973-060d-c549f93f4f6d" [ 802.110365] env[61985]: _type = "Task" [ 802.110365] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.119300] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b296a9-a9a8-b973-060d-c549f93f4f6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.265090] env[61985]: INFO nova.compute.manager [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Took 34.01 seconds to build instance. [ 802.350890] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935747, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.489906] env[61985]: DEBUG nova.network.neutron [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Successfully created port: b6ffda80-c267-47cb-92e7-6a930fbf0c27 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.512996] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 802.513307] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe3a4596-68f8-476f-b9a2-9c6a695d96d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.523313] env[61985]: DEBUG nova.compute.manager [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 802.526704] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 802.526704] env[61985]: value = "task-935748" [ 802.526704] env[61985]: _type = "Task" [ 802.526704] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.536249] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "refresh_cache-cfd59b61-cca9-48d5-85e1-1f45d13f1e88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.536441] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "refresh_cache-cfd59b61-cca9-48d5-85e1-1f45d13f1e88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.536561] env[61985]: DEBUG nova.network.neutron [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 802.542923] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935748, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.632117] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b296a9-a9a8-b973-060d-c549f93f4f6d, 'name': SearchDatastore_Task, 'duration_secs': 0.040806} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.637036] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.637036] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.637036] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.637036] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.637364] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.637364] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.637364] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.637687] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.637889] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.638149] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.638458] env[61985]: DEBUG nova.virt.hardware [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.638847] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.642025] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 4aca0ecb-4ae6-4400-accd-d71782b2806d/4aca0ecb-4ae6-4400-accd-d71782b2806d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 802.642025] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67c1504-ba77-452a-ac73-677617bd3af2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.647588] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8a6f848-b6f1-43e6-85bf-b7029d20e908 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.660817] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab034bb-7f83-47d1-be16-f3a5e3295d48 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.668514] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for the task: (returnval){ [ 802.668514] env[61985]: value = "task-935749" [ 802.668514] env[61985]: _type = "Task" [ 802.668514] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.691252] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 802.701068] env[61985]: DEBUG oslo.service.loopingcall [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.705499] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 802.706415] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b11ae826-456e-4680-a59a-77c202d09eed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.723608] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935749, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.731602] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.731602] env[61985]: value = "task-935750" [ 802.731602] env[61985]: _type = "Task" [ 802.731602] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.745219] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935750, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.767150] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6519b750-30e6-4bac-8fa7-97ec18a454cf tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "b659ff28-5101-4825-84ea-111351c81145" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.073s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.850630] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935747, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.045594] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7e9e6b-cd66-45ce-82a9-84de47e16b6c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.057496] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935748, 'name': PowerOffVM_Task, 'duration_secs': 0.193472} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.059665] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 803.059905] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 803.065022] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65bd780-958e-4205-b31b-ea12ab9be36e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.067061] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9cfa1f6-3ee4-4456-ac29-1ba1090cc83d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.106309] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 803.108219] env[61985]: DEBUG nova.network.neutron [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.110973] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1a401ec-62ff-4b28-a04f-c3d09905bcda {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.113349] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c777031f-83aa-46be-90f7-89d66435e172 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.122963] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167251ae-39ec-424b-9ec7-acf2ccb7720f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.142024] env[61985]: DEBUG nova.compute.provider_tree [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.187116] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935749, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.196296] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 803.196618] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 803.196737] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleting the datastore file [datastore1] f8bc8b71-0317-479d-b2f9-9471757f0774 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 803.197016] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a31a159-d55e-4548-be44-0b9f8b7b3b21 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.204016] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 803.204016] env[61985]: value = "task-935752" [ 803.204016] env[61985]: _type = "Task" [ 803.204016] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.214329] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935752, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.240894] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935750, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.271384] env[61985]: DEBUG nova.compute.manager [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 803.349567] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935747, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.406908] env[61985]: DEBUG nova.network.neutron [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Updating instance_info_cache with network_info: [{"id": "c529f7f6-f908-4a69-944f-b979f622f99f", "address": "fa:16:3e:20:1b:54", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc529f7f6-f9", "ovs_interfaceid": "c529f7f6-f908-4a69-944f-b979f622f99f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.549922] env[61985]: DEBUG nova.compute.manager [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 803.575162] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 803.575463] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 803.575638] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 803.575826] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 803.575968] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 803.576122] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 803.576341] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 803.576828] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 803.576828] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 803.576828] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 803.577020] env[61985]: DEBUG nova.virt.hardware [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 803.577878] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1750bf-3bd3-44a6-9403-b9865afa64fb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.585947] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daec3f17-c85f-491a-ad19-14286f024693 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.642681] env[61985]: DEBUG nova.scheduler.client.report [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.681573] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935749, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545036} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.681794] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 4aca0ecb-4ae6-4400-accd-d71782b2806d/4aca0ecb-4ae6-4400-accd-d71782b2806d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 803.682451] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.682451] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5c8b6d9-c16a-410b-8535-7da09e7dba8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.693243] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for the task: (returnval){ [ 803.693243] env[61985]: value = "task-935753" [ 803.693243] env[61985]: _type = "Task" [ 803.693243] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.699976] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935753, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.712532] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935752, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291201} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.716019] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 803.716019] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 803.716019] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 803.744936] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935750, 'name': CreateVM_Task, 'duration_secs': 0.536443} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.744936] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 803.744936] env[61985]: DEBUG oslo_concurrency.lockutils [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.744936] env[61985]: DEBUG oslo_concurrency.lockutils [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.744936] env[61985]: DEBUG oslo_concurrency.lockutils [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 803.744936] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d2d8e59-8002-4d90-b28f-8c8fc2e85fbb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.749367] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for the task: (returnval){ [ 803.749367] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5215bfb6-a146-7f5d-1022-fd552f0e9f68" [ 803.749367] env[61985]: _type = "Task" [ 803.749367] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.758766] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5215bfb6-a146-7f5d-1022-fd552f0e9f68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.804520] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.852760] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935747, 'name': ReconfigVM_Task, 'duration_secs': 1.858589} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.852760] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 761d5954-a5ca-4459-a1d6-bfc59b284bf4/761d5954-a5ca-4459-a1d6-bfc59b284bf4.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.852977] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b064772c-b9d5-4589-bc07-fd3922fb84d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.864210] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 803.864210] env[61985]: value = "task-935754" [ 803.864210] env[61985]: _type = "Task" [ 803.864210] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.868995] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935754, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.909501] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "refresh_cache-cfd59b61-cca9-48d5-85e1-1f45d13f1e88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.910614] env[61985]: DEBUG nova.compute.manager [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Instance network_info: |[{"id": "c529f7f6-f908-4a69-944f-b979f622f99f", "address": "fa:16:3e:20:1b:54", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc529f7f6-f9", "ovs_interfaceid": "c529f7f6-f908-4a69-944f-b979f622f99f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 803.910766] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:1b:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cd38fdec-d092-4a84-ab41-685f6dbb4f29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c529f7f6-f908-4a69-944f-b979f622f99f', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 803.924829] env[61985]: DEBUG oslo.service.loopingcall [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.925044] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 803.925208] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fffa9074-1506-46d4-9158-7fab54a1dc09 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.946652] env[61985]: DEBUG nova.compute.manager [req-b1f8dcd2-ad81-4413-bbce-11a174e28357 req-c2f1ca7f-c23a-4d54-9c64-401e804e7cab service nova] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Received event network-vif-plugged-b6ffda80-c267-47cb-92e7-6a930fbf0c27 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 803.947046] env[61985]: DEBUG oslo_concurrency.lockutils [req-b1f8dcd2-ad81-4413-bbce-11a174e28357 req-c2f1ca7f-c23a-4d54-9c64-401e804e7cab service nova] Acquiring lock "8f6cd002-b3c2-4276-b195-15d09a143d31-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.947134] env[61985]: DEBUG oslo_concurrency.lockutils [req-b1f8dcd2-ad81-4413-bbce-11a174e28357 req-c2f1ca7f-c23a-4d54-9c64-401e804e7cab service nova] Lock "8f6cd002-b3c2-4276-b195-15d09a143d31-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.947349] env[61985]: DEBUG oslo_concurrency.lockutils [req-b1f8dcd2-ad81-4413-bbce-11a174e28357 req-c2f1ca7f-c23a-4d54-9c64-401e804e7cab service nova] Lock "8f6cd002-b3c2-4276-b195-15d09a143d31-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.947540] env[61985]: DEBUG nova.compute.manager [req-b1f8dcd2-ad81-4413-bbce-11a174e28357 req-c2f1ca7f-c23a-4d54-9c64-401e804e7cab service nova] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] No waiting events found dispatching network-vif-plugged-b6ffda80-c267-47cb-92e7-6a930fbf0c27 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 803.947717] env[61985]: WARNING nova.compute.manager [req-b1f8dcd2-ad81-4413-bbce-11a174e28357 req-c2f1ca7f-c23a-4d54-9c64-401e804e7cab service nova] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Received unexpected event network-vif-plugged-b6ffda80-c267-47cb-92e7-6a930fbf0c27 for instance with vm_state building and task_state spawning. [ 803.949554] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 803.949554] env[61985]: value = "task-935755" [ 803.949554] env[61985]: _type = "Task" [ 803.949554] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.960052] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935755, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.995763] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "b659ff28-5101-4825-84ea-111351c81145" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.996016] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "b659ff28-5101-4825-84ea-111351c81145" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.996242] env[61985]: INFO nova.compute.manager [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Rebooting instance [ 804.041372] env[61985]: DEBUG nova.network.neutron [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Successfully updated port: b6ffda80-c267-47cb-92e7-6a930fbf0c27 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 804.081918] env[61985]: DEBUG nova.compute.manager [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Received event network-changed-c529f7f6-f908-4a69-944f-b979f622f99f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 804.082126] env[61985]: DEBUG nova.compute.manager [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Refreshing instance network info cache due to event network-changed-c529f7f6-f908-4a69-944f-b979f622f99f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 804.082371] env[61985]: DEBUG oslo_concurrency.lockutils [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] Acquiring lock "refresh_cache-cfd59b61-cca9-48d5-85e1-1f45d13f1e88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.082524] env[61985]: DEBUG oslo_concurrency.lockutils [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] Acquired lock "refresh_cache-cfd59b61-cca9-48d5-85e1-1f45d13f1e88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.082689] env[61985]: DEBUG nova.network.neutron [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Refreshing network info cache for port c529f7f6-f908-4a69-944f-b979f622f99f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 804.150845] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.646s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.151446] env[61985]: DEBUG nova.compute.manager [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 804.154741] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.757s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.154994] env[61985]: DEBUG nova.objects.instance [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lazy-loading 'resources' on Instance uuid 4add2608-8187-42a2-9897-3e5a9049757e {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.201661] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935753, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072839} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.202475] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.202796] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2979be4-c388-47c5-bd73-4a0c3e9de746 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.231178] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] 4aca0ecb-4ae6-4400-accd-d71782b2806d/4aca0ecb-4ae6-4400-accd-d71782b2806d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.231998] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-986cea32-60ec-4ad1-a7bc-447b2cc7baa0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.255910] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for the task: (returnval){ [ 804.255910] env[61985]: value = "task-935756" [ 804.255910] env[61985]: _type = "Task" [ 804.255910] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.263419] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5215bfb6-a146-7f5d-1022-fd552f0e9f68, 'name': SearchDatastore_Task, 'duration_secs': 0.009477} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.264248] env[61985]: DEBUG oslo_concurrency.lockutils [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.264523] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.264795] env[61985]: DEBUG oslo_concurrency.lockutils [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.264941] env[61985]: DEBUG oslo_concurrency.lockutils [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.265130] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.265768] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f46b365-3776-4904-a270-ce276e4283e9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.272159] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935756, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.278616] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.278852] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 804.279929] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe7190ac-c9ce-4914-83d6-4ab9d1704a1d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.286117] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for the task: (returnval){ [ 804.286117] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b8ef49-c50c-7ebd-3ced-29e282c53a8f" [ 804.286117] env[61985]: _type = "Task" [ 804.286117] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.294701] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b8ef49-c50c-7ebd-3ced-29e282c53a8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.372441] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935754, 'name': Rename_Task, 'duration_secs': 0.14468} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.372859] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 804.373194] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72c22b48-7295-42e9-b3c6-1122707d7262 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.379660] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 804.379660] env[61985]: value = "task-935757" [ 804.379660] env[61985]: _type = "Task" [ 804.379660] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.388671] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935757, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.462900] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935755, 'name': CreateVM_Task, 'duration_secs': 0.443958} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.463194] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 804.463976] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.464244] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.464638] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 804.465157] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dbaf5ff-2dcb-4dc5-a6ca-50c1eb97816e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.477020] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 804.477020] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52eb5133-86a3-5afd-d7fe-d0b6a5a9820e" [ 804.477020] env[61985]: _type = "Task" [ 804.477020] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.488511] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52eb5133-86a3-5afd-d7fe-d0b6a5a9820e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.525687] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.529324] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquired lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.529550] env[61985]: DEBUG nova.network.neutron [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.545248] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "refresh_cache-8f6cd002-b3c2-4276-b195-15d09a143d31" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.545248] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "refresh_cache-8f6cd002-b3c2-4276-b195-15d09a143d31" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.545248] env[61985]: DEBUG nova.network.neutron [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.658389] env[61985]: DEBUG nova.compute.utils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.663769] env[61985]: DEBUG nova.compute.manager [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 804.663769] env[61985]: DEBUG nova.network.neutron [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 804.766468] env[61985]: DEBUG nova.policy [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88a11f4b865a4d92a8d5e7d238363519', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8cf6728cd4fa40119f91548ed7a717b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 804.775825] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935756, 'name': ReconfigVM_Task, 'duration_secs': 0.274188} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.778101] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.778679] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.778679] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.778679] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.778975] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.778975] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.779210] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.779409] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.779584] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.779746] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.779917] env[61985]: DEBUG nova.virt.hardware [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.783015] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Reconfigured VM instance instance-0000002e to attach disk [datastore1] 4aca0ecb-4ae6-4400-accd-d71782b2806d/4aca0ecb-4ae6-4400-accd-d71782b2806d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.783015] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57cf3718-5672-4630-9689-dd867483ec11 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.786153] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5068e500-4ad0-4510-a84f-5a614743f851 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.798658] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b4e20c-da4a-4fc7-8b9b-494d410d5d83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.806976] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for the task: (returnval){ [ 804.806976] env[61985]: value = "task-935758" [ 804.806976] env[61985]: _type = "Task" [ 804.806976] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.807675] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b8ef49-c50c-7ebd-3ced-29e282c53a8f, 'name': SearchDatastore_Task, 'duration_secs': 0.008991} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.813798] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0b021d3-0c81-4792-969f-2e82fdcd52c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.827215] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:b2:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bff6c3a1-cc80-46ca-86c0-6dbb029edddb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '787b8b7e-dea3-41b7-9349-4f87afe7b88d', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.835252] env[61985]: DEBUG oslo.service.loopingcall [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.838717] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 804.839480] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c79a739-3bd5-415c-ac7c-6226ddf14292 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.862840] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935758, 'name': Rename_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.863169] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for the task: (returnval){ [ 804.863169] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cf07b7-ce09-f8b5-5290-651cee2fe49e" [ 804.863169] env[61985]: _type = "Task" [ 804.863169] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.867861] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.867861] env[61985]: value = "task-935759" [ 804.867861] env[61985]: _type = "Task" [ 804.867861] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.875389] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cf07b7-ce09-f8b5-5290-651cee2fe49e, 'name': SearchDatastore_Task, 'duration_secs': 0.008664} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.876075] env[61985]: DEBUG oslo_concurrency.lockutils [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.876414] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 804.877368] env[61985]: DEBUG nova.network.neutron [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Updated VIF entry in instance network info cache for port c529f7f6-f908-4a69-944f-b979f622f99f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 804.877975] env[61985]: DEBUG nova.network.neutron [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Updating instance_info_cache with network_info: [{"id": "c529f7f6-f908-4a69-944f-b979f622f99f", "address": "fa:16:3e:20:1b:54", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc529f7f6-f9", "ovs_interfaceid": "c529f7f6-f908-4a69-944f-b979f622f99f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.881342] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-561d9057-43d8-4470-96d9-fcf2067cf792 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.890055] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935759, 'name': CreateVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.897638] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935757, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.901498] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for the task: (returnval){ [ 804.901498] env[61985]: value = "task-935760" [ 804.901498] env[61985]: _type = "Task" [ 804.901498] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.912194] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935760, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.986776] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52eb5133-86a3-5afd-d7fe-d0b6a5a9820e, 'name': SearchDatastore_Task, 'duration_secs': 0.012844} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.987101] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.987430] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.987724] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.987885] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.988077] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.988357] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dfe2050c-97ac-44f9-bd33-d97e47652bea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.998644] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.998644] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 804.999314] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6566dc98-3c8a-436d-bfbf-a46322fe54c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.006556] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 805.006556] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b109c3-8f47-0d4a-0779-b8e5cbaf1111" [ 805.006556] env[61985]: _type = "Task" [ 805.006556] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.014481] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b109c3-8f47-0d4a-0779-b8e5cbaf1111, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.123010] env[61985]: DEBUG nova.network.neutron [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 805.127185] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8498211a-8680-414c-8b0f-8717eaeccab3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.138201] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f1de17-ac45-489c-b18a-0fe531de183a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.174116] env[61985]: DEBUG nova.compute.manager [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 805.180660] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cda9591f-8a53-4945-9dd3-65cb4bc7a438 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.190388] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c15b794-e2d4-4dbd-bf53-8381f26533ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.206192] env[61985]: DEBUG nova.compute.provider_tree [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.309697] env[61985]: DEBUG nova.network.neutron [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Updating instance_info_cache with network_info: [{"id": "bf071457-a423-45ed-9a6e-22e52c290be8", "address": "fa:16:3e:8b:24:a0", "network": {"id": "a3e42a35-4a68-42be-bd75-5bf48aeab85b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1220560148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e8e226d5c9240928ccc35013189235c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf071457-a4", "ovs_interfaceid": "bf071457-a423-45ed-9a6e-22e52c290be8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.332520] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935758, 'name': Rename_Task, 'duration_secs': 0.157112} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.332520] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 805.332520] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f0804304-0983-4817-abc5-e497eb84d68f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.338118] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for the task: (returnval){ [ 805.338118] env[61985]: value = "task-935761" [ 805.338118] env[61985]: _type = "Task" [ 805.338118] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.349168] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935761, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.382516] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935759, 'name': CreateVM_Task, 'duration_secs': 0.508244} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.382516] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 805.382516] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.382516] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.382516] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.382516] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44c6006b-5ecb-4a97-88c5-221e701991aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.385069] env[61985]: DEBUG oslo_concurrency.lockutils [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] Releasing lock "refresh_cache-cfd59b61-cca9-48d5-85e1-1f45d13f1e88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.385069] env[61985]: DEBUG nova.compute.manager [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Received event network-changed-bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 805.385526] env[61985]: DEBUG nova.compute.manager [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Refreshing instance network info cache due to event network-changed-bf071457-a423-45ed-9a6e-22e52c290be8. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 805.385526] env[61985]: DEBUG oslo_concurrency.lockutils [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] Acquiring lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.396913] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935757, 'name': PowerOnVM_Task} progress is 37%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.397283] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 805.397283] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528f004a-5101-1fef-1906-752e952a0e77" [ 805.397283] env[61985]: _type = "Task" [ 805.397283] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.410071] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528f004a-5101-1fef-1906-752e952a0e77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.418464] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935760, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.516930] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b109c3-8f47-0d4a-0779-b8e5cbaf1111, 'name': SearchDatastore_Task, 'duration_secs': 0.009616} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.517902] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1a1b902-3dc0-432c-9f80-48f53e7053b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.522738] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 805.522738] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5270c9a6-1477-eca7-dc8f-41827f081263" [ 805.522738] env[61985]: _type = "Task" [ 805.522738] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.530113] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5270c9a6-1477-eca7-dc8f-41827f081263, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.651304] env[61985]: DEBUG nova.network.neutron [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Successfully created port: 5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 805.710627] env[61985]: DEBUG nova.scheduler.client.report [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.714717] env[61985]: DEBUG nova.network.neutron [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Updating instance_info_cache with network_info: [{"id": "b6ffda80-c267-47cb-92e7-6a930fbf0c27", "address": "fa:16:3e:30:3b:08", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6ffda80-c2", "ovs_interfaceid": "b6ffda80-c267-47cb-92e7-6a930fbf0c27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.816944] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Releasing lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.819105] env[61985]: DEBUG oslo_concurrency.lockutils [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] Acquired lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.819349] env[61985]: DEBUG nova.network.neutron [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Refreshing network info cache for port bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 805.820561] env[61985]: DEBUG nova.compute.manager [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.821402] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45165702-4b3b-4e1d-a7c4-ded2693b2ae1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.851449] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935761, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.896810] env[61985]: DEBUG oslo_vmware.api [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935757, 'name': PowerOnVM_Task, 'duration_secs': 1.454901} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.897387] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 805.897640] env[61985]: INFO nova.compute.manager [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Took 10.07 seconds to spawn the instance on the hypervisor. [ 805.899250] env[61985]: DEBUG nova.compute.manager [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.900106] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74c20bb-7989-4935-85e7-508e5c6e0397 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.923836] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528f004a-5101-1fef-1906-752e952a0e77, 'name': SearchDatastore_Task, 'duration_secs': 0.030474} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.924142] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.924392] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.924608] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.924752] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.924951] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.928769] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5478d54-9d31-4d92-884d-120153aba563 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.930939] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935760, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.580704} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.931234] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 805.931513] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.932271] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc5944ca-bb90-4986-8015-7b849ed52356 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.938614] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.938805] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 805.942309] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8433e967-4c31-4585-93c5-d86ebe5ad5cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.944026] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for the task: (returnval){ [ 805.944026] env[61985]: value = "task-935762" [ 805.944026] env[61985]: _type = "Task" [ 805.944026] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.948491] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 805.948491] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b01d7e-368b-4940-815a-4c3be8081326" [ 805.948491] env[61985]: _type = "Task" [ 805.948491] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.956995] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935762, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.962314] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b01d7e-368b-4940-815a-4c3be8081326, 'name': SearchDatastore_Task, 'duration_secs': 0.012419} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.963278] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a9f9ecc-3b24-4b9c-80bf-0e07408333cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.968926] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 805.968926] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e8b2e6-59af-0f81-fdf1-2f5c8689d98e" [ 805.968926] env[61985]: _type = "Task" [ 805.968926] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.978522] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e8b2e6-59af-0f81-fdf1-2f5c8689d98e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.034620] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5270c9a6-1477-eca7-dc8f-41827f081263, 'name': SearchDatastore_Task, 'duration_secs': 0.009735} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.034713] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.035015] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] cfd59b61-cca9-48d5-85e1-1f45d13f1e88/cfd59b61-cca9-48d5-85e1-1f45d13f1e88.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 806.035282] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d9ec95bf-02cc-48ca-b3d2-b587a20c5bcf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.042070] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 806.042070] env[61985]: value = "task-935763" [ 806.042070] env[61985]: _type = "Task" [ 806.042070] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.050935] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935763, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.191106] env[61985]: DEBUG nova.compute.manager [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 806.213267] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.213521] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.213679] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.213863] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.214010] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.214229] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.214445] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.214601] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.214767] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.214929] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.215112] env[61985]: DEBUG nova.virt.hardware [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.215959] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77702004-635a-4bf0-bc85-61660538b079 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.219148] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.065s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.221681] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "refresh_cache-8f6cd002-b3c2-4276-b195-15d09a143d31" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.222160] env[61985]: DEBUG nova.compute.manager [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Instance network_info: |[{"id": "b6ffda80-c267-47cb-92e7-6a930fbf0c27", "address": "fa:16:3e:30:3b:08", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6ffda80-c2", "ovs_interfaceid": "b6ffda80-c267-47cb-92e7-6a930fbf0c27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 806.222667] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.712s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.224292] env[61985]: INFO nova.compute.claims [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.227464] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:3b:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7908211b-df93-467b-87a8-3c3d29b03de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6ffda80-c267-47cb-92e7-6a930fbf0c27', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 806.235865] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Creating folder: Project (5625013599b44418bd56eb604e14be58). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 806.236560] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4709ea0-1071-4f4b-ae13-06882fed51fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.247132] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfcfb4f-9e55-4b02-9b81-bc2bb030c445 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.251034] env[61985]: INFO nova.scheduler.client.report [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Deleted allocations for instance 4add2608-8187-42a2-9897-3e5a9049757e [ 806.255492] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Created folder: Project (5625013599b44418bd56eb604e14be58) in parent group-v211285. [ 806.255715] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Creating folder: Instances. Parent ref: group-v211332. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 806.260184] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0fb5b13c-0b79-4968-a32d-22370ec20119 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.278053] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Created folder: Instances in parent group-v211332. [ 806.278280] env[61985]: DEBUG oslo.service.loopingcall [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.278474] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 806.278711] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9662f1f5-e0cc-41f7-b4d3-c7b0a9eb660f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.300864] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 806.300864] env[61985]: value = "task-935766" [ 806.300864] env[61985]: _type = "Task" [ 806.300864] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.310335] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935766, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.328074] env[61985]: DEBUG nova.compute.manager [req-2e257009-9443-4892-9ae1-891068936b74 req-b2a43f0a-8de4-4f1c-8b91-63a3e9baf7b1 service nova] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Received event network-changed-b6ffda80-c267-47cb-92e7-6a930fbf0c27 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 806.328369] env[61985]: DEBUG nova.compute.manager [req-2e257009-9443-4892-9ae1-891068936b74 req-b2a43f0a-8de4-4f1c-8b91-63a3e9baf7b1 service nova] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Refreshing instance network info cache due to event network-changed-b6ffda80-c267-47cb-92e7-6a930fbf0c27. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 806.328627] env[61985]: DEBUG oslo_concurrency.lockutils [req-2e257009-9443-4892-9ae1-891068936b74 req-b2a43f0a-8de4-4f1c-8b91-63a3e9baf7b1 service nova] Acquiring lock "refresh_cache-8f6cd002-b3c2-4276-b195-15d09a143d31" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.328838] env[61985]: DEBUG oslo_concurrency.lockutils [req-2e257009-9443-4892-9ae1-891068936b74 req-b2a43f0a-8de4-4f1c-8b91-63a3e9baf7b1 service nova] Acquired lock "refresh_cache-8f6cd002-b3c2-4276-b195-15d09a143d31" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.329642] env[61985]: DEBUG nova.network.neutron [req-2e257009-9443-4892-9ae1-891068936b74 req-b2a43f0a-8de4-4f1c-8b91-63a3e9baf7b1 service nova] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Refreshing network info cache for port b6ffda80-c267-47cb-92e7-6a930fbf0c27 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 806.350491] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935761, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.436492] env[61985]: INFO nova.compute.manager [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Took 34.02 seconds to build instance. [ 806.454480] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935762, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121875} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.454697] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 806.455557] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b78914-f794-4ca7-832a-d4da96afdaf6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.477528] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Reconfiguring VM instance instance-00000028 to attach disk [datastore1] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 806.480406] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1923c4ce-c4e7-4583-925f-806b3f9d459f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.504613] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e8b2e6-59af-0f81-fdf1-2f5c8689d98e, 'name': SearchDatastore_Task, 'duration_secs': 0.01637} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.505913] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.506265] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] f8bc8b71-0317-479d-b2f9-9471757f0774/f8bc8b71-0317-479d-b2f9-9471757f0774.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 806.507083] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for the task: (returnval){ [ 806.507083] env[61985]: value = "task-935767" [ 806.507083] env[61985]: _type = "Task" [ 806.507083] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.507083] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4a6b92e-8f17-4c1f-8af2-3cdaf9e7e89d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.517645] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935767, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.518923] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 806.518923] env[61985]: value = "task-935768" [ 806.518923] env[61985]: _type = "Task" [ 806.518923] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.529226] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935768, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.555114] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935763, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.765425] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f5e54cf4-7943-4841-b09e-981c4917a581 tempest-VolumesAdminNegativeTest-527673646 tempest-VolumesAdminNegativeTest-527673646-project-member] Lock "4add2608-8187-42a2-9897-3e5a9049757e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.448s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.812467] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935766, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.835903] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ccf424b-dd42-439b-94b8-9b28c1145674 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.844996] env[61985]: DEBUG nova.network.neutron [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Updated VIF entry in instance network info cache for port bf071457-a423-45ed-9a6e-22e52c290be8. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 806.845536] env[61985]: DEBUG nova.network.neutron [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Updating instance_info_cache with network_info: [{"id": "bf071457-a423-45ed-9a6e-22e52c290be8", "address": "fa:16:3e:8b:24:a0", "network": {"id": "a3e42a35-4a68-42be-bd75-5bf48aeab85b", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1220560148-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e8e226d5c9240928ccc35013189235c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf071457-a4", "ovs_interfaceid": "bf071457-a423-45ed-9a6e-22e52c290be8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.849532] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Doing hard reboot of VM {{(pid=61985) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 806.849974] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-dc010edd-b22e-4246-a251-fd76e021356e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.855210] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935761, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.867625] env[61985]: DEBUG oslo_vmware.api [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 806.867625] env[61985]: value = "task-935769" [ 806.867625] env[61985]: _type = "Task" [ 806.867625] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.879137] env[61985]: DEBUG oslo_vmware.api [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935769, 'name': ResetVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.939802] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17537e22-09f2-4e7e-af9f-f4602a109030 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.662s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.024683] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935767, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.033738] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935768, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.054245] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935763, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.84256} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.054636] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] cfd59b61-cca9-48d5-85e1-1f45d13f1e88/cfd59b61-cca9-48d5-85e1-1f45d13f1e88.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 807.054899] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.055206] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18ca895c-51b4-427b-85ad-1829056021c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.063729] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 807.063729] env[61985]: value = "task-935770" [ 807.063729] env[61985]: _type = "Task" [ 807.063729] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.072838] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935770, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.173170] env[61985]: DEBUG nova.network.neutron [req-2e257009-9443-4892-9ae1-891068936b74 req-b2a43f0a-8de4-4f1c-8b91-63a3e9baf7b1 service nova] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Updated VIF entry in instance network info cache for port b6ffda80-c267-47cb-92e7-6a930fbf0c27. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 807.173908] env[61985]: DEBUG nova.network.neutron [req-2e257009-9443-4892-9ae1-891068936b74 req-b2a43f0a-8de4-4f1c-8b91-63a3e9baf7b1 service nova] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Updating instance_info_cache with network_info: [{"id": "b6ffda80-c267-47cb-92e7-6a930fbf0c27", "address": "fa:16:3e:30:3b:08", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6ffda80-c2", "ovs_interfaceid": "b6ffda80-c267-47cb-92e7-6a930fbf0c27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.312724] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935766, 'name': CreateVM_Task, 'duration_secs': 0.699743} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.312724] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 807.313451] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.313620] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.313958] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.314234] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccc4601e-71ce-4b59-aa4c-6f6e33a83b7e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.325239] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 807.325239] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5230a8f9-b22e-ae55-1e0a-b1b9919ab03a" [ 807.325239] env[61985]: _type = "Task" [ 807.325239] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.335143] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5230a8f9-b22e-ae55-1e0a-b1b9919ab03a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.348237] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935761, 'name': PowerOnVM_Task} progress is 82%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.351225] env[61985]: DEBUG oslo_concurrency.lockutils [req-95fb2063-d82b-4509-b53a-36b3aa316311 req-6db3d0b6-e644-4401-ab48-3b5416b9cd68 service nova] Releasing lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.377087] env[61985]: DEBUG oslo_vmware.api [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935769, 'name': ResetVM_Task, 'duration_secs': 0.095035} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.377373] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Did hard reboot of VM {{(pid=61985) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 807.377563] env[61985]: DEBUG nova.compute.manager [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.379831] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc9cf2f-4e4e-4b29-ac02-4ed2bec20564 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.444669] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 807.520497] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935767, 'name': ReconfigVM_Task, 'duration_secs': 0.762396} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.522430] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Reconfigured VM instance instance-00000028 to attach disk [datastore1] 794cff68-6e26-4607-96f7-eaeb41182551/794cff68-6e26-4607-96f7-eaeb41182551.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 807.523364] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8961fc7-f586-4e11-a84d-73a83e10c1b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.533585] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935768, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.81322} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.536836] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] f8bc8b71-0317-479d-b2f9-9471757f0774/f8bc8b71-0317-479d-b2f9-9471757f0774.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 807.537068] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.537394] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for the task: (returnval){ [ 807.537394] env[61985]: value = "task-935771" [ 807.537394] env[61985]: _type = "Task" [ 807.537394] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.537974] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8db1fc56-493b-4631-9c3c-e8f11df0f0c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.548360] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 807.548360] env[61985]: value = "task-935772" [ 807.548360] env[61985]: _type = "Task" [ 807.548360] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.559796] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935772, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.575614] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935770, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.166053} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.575838] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.576814] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df5bc67-52d3-4237-af5d-996679b94c6a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.610936] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] cfd59b61-cca9-48d5-85e1-1f45d13f1e88/cfd59b61-cca9-48d5-85e1-1f45d13f1e88.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.613874] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f818f30-5390-4477-bc1d-04ba9afb625d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.633901] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 807.633901] env[61985]: value = "task-935773" [ 807.633901] env[61985]: _type = "Task" [ 807.633901] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.642664] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935773, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.677748] env[61985]: DEBUG oslo_concurrency.lockutils [req-2e257009-9443-4892-9ae1-891068936b74 req-b2a43f0a-8de4-4f1c-8b91-63a3e9baf7b1 service nova] Releasing lock "refresh_cache-8f6cd002-b3c2-4276-b195-15d09a143d31" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.731922] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd341f0f-a324-42f0-9a71-d1213144b342 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.740576] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88de46e-68f1-4114-9622-a820fdc2b614 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.771921] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f95f23-96d5-426b-a2bb-74300a084239 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.779907] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1422bf1f-b093-4953-90ee-43f9b4e38b92 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.794745] env[61985]: DEBUG nova.compute.provider_tree [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.811211] env[61985]: DEBUG nova.compute.manager [req-90a272aa-6598-4a15-b5df-a746893ea083 req-8cf69ba4-2787-4b03-a4cb-17b277cc0702 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Received event network-vif-plugged-5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 807.811459] env[61985]: DEBUG oslo_concurrency.lockutils [req-90a272aa-6598-4a15-b5df-a746893ea083 req-8cf69ba4-2787-4b03-a4cb-17b277cc0702 service nova] Acquiring lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.811722] env[61985]: DEBUG oslo_concurrency.lockutils [req-90a272aa-6598-4a15-b5df-a746893ea083 req-8cf69ba4-2787-4b03-a4cb-17b277cc0702 service nova] Lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.811946] env[61985]: DEBUG oslo_concurrency.lockutils [req-90a272aa-6598-4a15-b5df-a746893ea083 req-8cf69ba4-2787-4b03-a4cb-17b277cc0702 service nova] Lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.812273] env[61985]: DEBUG nova.compute.manager [req-90a272aa-6598-4a15-b5df-a746893ea083 req-8cf69ba4-2787-4b03-a4cb-17b277cc0702 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] No waiting events found dispatching network-vif-plugged-5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 807.812620] env[61985]: WARNING nova.compute.manager [req-90a272aa-6598-4a15-b5df-a746893ea083 req-8cf69ba4-2787-4b03-a4cb-17b277cc0702 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Received unexpected event network-vif-plugged-5aa010f5-afab-4ea2-80c4-0d0626458099 for instance with vm_state building and task_state spawning. [ 807.835812] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5230a8f9-b22e-ae55-1e0a-b1b9919ab03a, 'name': SearchDatastore_Task, 'duration_secs': 0.025884} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.836351] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.836640] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.837136] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.837348] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.837571] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.838044] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a47934cd-0016-4276-b802-528fe1a2c742 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.850705] env[61985]: DEBUG oslo_vmware.api [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935761, 'name': PowerOnVM_Task, 'duration_secs': 2.081249} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.850705] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 807.851396] env[61985]: INFO nova.compute.manager [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Took 9.48 seconds to spawn the instance on the hypervisor. [ 807.851630] env[61985]: DEBUG nova.compute.manager [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.852575] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1848229b-0c7d-4b62-9999-01e23cd1d70d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.860430] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.860630] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 807.863433] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-babaceb7-d279-458c-ad0c-239ae0466a04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.874654] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 807.874654] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526307d5-42d6-a766-cb2a-9152b23e71d3" [ 807.874654] env[61985]: _type = "Task" [ 807.874654] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.884758] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526307d5-42d6-a766-cb2a-9152b23e71d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.894497] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b3c294c-0f70-4c62-b6dd-447db8967ac3 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "b659ff28-5101-4825-84ea-111351c81145" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.898s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.969144] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.975920] env[61985]: DEBUG nova.network.neutron [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Successfully updated port: 5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 808.049636] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935771, 'name': Rename_Task, 'duration_secs': 0.441555} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.049948] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 808.050237] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ae57822-34bc-43eb-810b-208a6fe413b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.060013] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935772, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067458} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.061172] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.061508] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Waiting for the task: (returnval){ [ 808.061508] env[61985]: value = "task-935774" [ 808.061508] env[61985]: _type = "Task" [ 808.061508] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.062175] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298fa2c2-99ad-42d5-850d-86f9a500359d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.072016] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.089561] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] f8bc8b71-0317-479d-b2f9-9471757f0774/f8bc8b71-0317-479d-b2f9-9471757f0774.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.089844] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f572c0c-8672-4ae8-a2a7-458fa0b2497f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.110887] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 808.110887] env[61985]: value = "task-935775" [ 808.110887] env[61985]: _type = "Task" [ 808.110887] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.122706] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935775, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.142724] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935773, 'name': ReconfigVM_Task, 'duration_secs': 0.290778} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.143019] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Reconfigured VM instance instance-0000002f to attach disk [datastore1] cfd59b61-cca9-48d5-85e1-1f45d13f1e88/cfd59b61-cca9-48d5-85e1-1f45d13f1e88.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.143829] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e9f8cf6-11eb-4187-ba0a-098af3ccf14e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.149582] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 808.149582] env[61985]: value = "task-935776" [ 808.149582] env[61985]: _type = "Task" [ 808.149582] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.162097] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935776, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.299330] env[61985]: DEBUG nova.scheduler.client.report [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.385505] env[61985]: INFO nova.compute.manager [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Took 29.17 seconds to build instance. [ 808.393530] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526307d5-42d6-a766-cb2a-9152b23e71d3, 'name': SearchDatastore_Task, 'duration_secs': 0.037116} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.394680] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09fe55bf-494e-4b61-973d-c590b941564e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.402062] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 808.402062] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52efbd57-96ae-b886-57b4-24629be23d73" [ 808.402062] env[61985]: _type = "Task" [ 808.402062] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.412030] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52efbd57-96ae-b886-57b4-24629be23d73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.479103] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquiring lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.479280] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquired lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.479437] env[61985]: DEBUG nova.network.neutron [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 808.574844] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935774, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.623521] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935775, 'name': ReconfigVM_Task, 'duration_secs': 0.449864} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.623853] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Reconfigured VM instance instance-0000002b to attach disk [datastore2] f8bc8b71-0317-479d-b2f9-9471757f0774/f8bc8b71-0317-479d-b2f9-9471757f0774.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.624498] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0cce02cc-e93e-4c15-babb-42b21bca72cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.632354] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 808.632354] env[61985]: value = "task-935777" [ 808.632354] env[61985]: _type = "Task" [ 808.632354] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.640910] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935777, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.659552] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935776, 'name': Rename_Task, 'duration_secs': 0.209814} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.659903] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 808.660162] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-344c1052-8d43-4718-9836-08ffcc1bd33f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.666752] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 808.666752] env[61985]: value = "task-935778" [ 808.666752] env[61985]: _type = "Task" [ 808.666752] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.674743] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935778, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.806491] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.584s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.807179] env[61985]: DEBUG nova.compute.manager [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 808.810688] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 19.139s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.887023] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d87afff-18f0-4afa-9fc6-53fb08a82af1 tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lock "4aca0ecb-4ae6-4400-accd-d71782b2806d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.540s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.913270] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52efbd57-96ae-b886-57b4-24629be23d73, 'name': SearchDatastore_Task, 'duration_secs': 0.011407} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.914130] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.914130] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8f6cd002-b3c2-4276-b195-15d09a143d31/8f6cd002-b3c2-4276-b195-15d09a143d31.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 808.914342] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e4dd054-42eb-4d12-b47c-e24476f09e53 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.921579] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 808.921579] env[61985]: value = "task-935779" [ 808.921579] env[61985]: _type = "Task" [ 808.921579] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.932405] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935779, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.076559] env[61985]: DEBUG oslo_vmware.api [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Task: {'id': task-935774, 'name': PowerOnVM_Task, 'duration_secs': 0.622594} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.076795] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 809.077039] env[61985]: DEBUG nova.compute.manager [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 809.077876] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3cf250d-fe31-435b-9979-bc74d6b79c41 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.103547] env[61985]: DEBUG nova.network.neutron [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 809.151315] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935777, 'name': Rename_Task, 'duration_secs': 0.159735} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.151692] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 809.151955] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7aaa8b3e-1ba4-4d78-bef2-da4573aab553 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.160634] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 809.160634] env[61985]: value = "task-935780" [ 809.160634] env[61985]: _type = "Task" [ 809.160634] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.170388] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.179914] env[61985]: DEBUG oslo_vmware.api [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935778, 'name': PowerOnVM_Task, 'duration_secs': 0.469239} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.180773] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 809.181036] env[61985]: INFO nova.compute.manager [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Took 8.23 seconds to spawn the instance on the hypervisor. [ 809.181264] env[61985]: DEBUG nova.compute.manager [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 809.182167] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0866bf-d141-4631-9d85-2b2610a1de70 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.319453] env[61985]: DEBUG nova.compute.utils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.323482] env[61985]: INFO nova.compute.claims [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.331960] env[61985]: DEBUG nova.compute.manager [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 809.332166] env[61985]: DEBUG nova.network.neutron [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 809.395685] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.408776] env[61985]: DEBUG nova.network.neutron [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Updating instance_info_cache with network_info: [{"id": "5aa010f5-afab-4ea2-80c4-0d0626458099", "address": "fa:16:3e:67:8d:3f", "network": {"id": "1ccd4f44-07f3-492a-a774-e9dfbee72280", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1344902357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cf6728cd4fa40119f91548ed7a717b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "779b8e65-8b9e-427e-af08-910febd65bfa", "external-id": "nsx-vlan-transportzone-906", "segmentation_id": 906, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa010f5-af", "ovs_interfaceid": "5aa010f5-afab-4ea2-80c4-0d0626458099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.420510] env[61985]: DEBUG nova.policy [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73b91a8d47584100813085e570f59ca2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd1b149ad11914d118f261481314b5a9d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 809.438167] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935779, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.605228] env[61985]: DEBUG oslo_concurrency.lockutils [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.675369] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935780, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.708602] env[61985]: INFO nova.compute.manager [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Took 27.23 seconds to build instance. [ 809.829164] env[61985]: DEBUG nova.compute.manager [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 809.838367] env[61985]: INFO nova.compute.resource_tracker [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating resource usage from migration feff90ec-33ae-4a19-ab75-42f6fbb33485 [ 809.911569] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Releasing lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.911906] env[61985]: DEBUG nova.compute.manager [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Instance network_info: |[{"id": "5aa010f5-afab-4ea2-80c4-0d0626458099", "address": "fa:16:3e:67:8d:3f", "network": {"id": "1ccd4f44-07f3-492a-a774-e9dfbee72280", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1344902357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cf6728cd4fa40119f91548ed7a717b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "779b8e65-8b9e-427e-af08-910febd65bfa", "external-id": "nsx-vlan-transportzone-906", "segmentation_id": 906, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa010f5-af", "ovs_interfaceid": "5aa010f5-afab-4ea2-80c4-0d0626458099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 809.916020] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:8d:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '779b8e65-8b9e-427e-af08-910febd65bfa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5aa010f5-afab-4ea2-80c4-0d0626458099', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.924463] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Creating folder: Project (8cf6728cd4fa40119f91548ed7a717b6). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 809.925589] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.928364] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d6d1a0e-b35a-4a3b-87f8-3e5077314e16 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.938819] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935779, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690083} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.939118] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8f6cd002-b3c2-4276-b195-15d09a143d31/8f6cd002-b3c2-4276-b195-15d09a143d31.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 809.939342] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.939586] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d764e2e7-0e83-4575-8ed5-b0f8b72e7fe8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.943102] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Created folder: Project (8cf6728cd4fa40119f91548ed7a717b6) in parent group-v211285. [ 809.943248] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Creating folder: Instances. Parent ref: group-v211335. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 809.944092] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5488c00c-e15e-4822-98fb-f42c4b5bae7b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.949834] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 809.949834] env[61985]: value = "task-935782" [ 809.949834] env[61985]: _type = "Task" [ 809.949834] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.956554] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Created folder: Instances in parent group-v211335. [ 809.956910] env[61985]: DEBUG oslo.service.loopingcall [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.959928] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 809.960223] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935782, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.960437] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13876da1-3948-434b-a3ed-28bd2dac6dfa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.983667] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.983667] env[61985]: value = "task-935784" [ 809.983667] env[61985]: _type = "Task" [ 809.983667] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.002019] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935784, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.023555] env[61985]: DEBUG nova.network.neutron [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Successfully created port: f734b929-a7bc-4469-b204-3b5afd314844 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.173192] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935780, 'name': PowerOnVM_Task} progress is 37%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.210940] env[61985]: DEBUG oslo_concurrency.lockutils [None req-283c7c22-4401-43f2-b3ad-e8c11a0ecfe9 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.399s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.266063] env[61985]: DEBUG nova.compute.manager [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Received event network-changed-5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 810.266154] env[61985]: DEBUG nova.compute.manager [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Refreshing instance network info cache due to event network-changed-5aa010f5-afab-4ea2-80c4-0d0626458099. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 810.266307] env[61985]: DEBUG oslo_concurrency.lockutils [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] Acquiring lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.266443] env[61985]: DEBUG oslo_concurrency.lockutils [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] Acquired lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.266631] env[61985]: DEBUG nova.network.neutron [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Refreshing network info cache for port 5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 810.389703] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b702c8f-3398-4d77-a62a-0bf94eb4bdc6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.398216] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb71ad2-7c7e-412d-95aa-9a20dce348c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.429887] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fb7155-c2f3-436d-8e35-4ac26f8b9e17 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.437719] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733c2cfc-a617-4cc8-a1d0-259a3cec0720 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.452698] env[61985]: DEBUG nova.compute.provider_tree [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.463688] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935782, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110722} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.463935] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 810.464809] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cd87c7-77dd-4d1a-81af-ffbb1212c212 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.493940] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 8f6cd002-b3c2-4276-b195-15d09a143d31/8f6cd002-b3c2-4276-b195-15d09a143d31.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.494903] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a66a9d05-5f93-4748-beb8-20aa71d1b153 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.520024] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 810.520024] env[61985]: value = "task-935785" [ 810.520024] env[61985]: _type = "Task" [ 810.520024] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.523043] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935784, 'name': CreateVM_Task, 'duration_secs': 0.529998} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.524910] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 810.525721] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.526323] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.526323] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 810.527972] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abea9731-1f2c-4808-8039-36f5a762e3ed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.533308] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935785, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.536778] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for the task: (returnval){ [ 810.536778] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52650b8d-b3b4-fcbd-6e67-cbf7a415b702" [ 810.536778] env[61985]: _type = "Task" [ 810.536778] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.545599] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52650b8d-b3b4-fcbd-6e67-cbf7a415b702, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.662092] env[61985]: INFO nova.compute.manager [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Rescuing [ 810.662410] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "refresh_cache-cfd59b61-cca9-48d5-85e1-1f45d13f1e88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.662596] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "refresh_cache-cfd59b61-cca9-48d5-85e1-1f45d13f1e88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.662794] env[61985]: DEBUG nova.network.neutron [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 810.675359] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935780, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.714036] env[61985]: DEBUG nova.compute.manager [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.824980] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "b659ff28-5101-4825-84ea-111351c81145" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.825357] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "b659ff28-5101-4825-84ea-111351c81145" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.825787] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "b659ff28-5101-4825-84ea-111351c81145-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.825949] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "b659ff28-5101-4825-84ea-111351c81145-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.826198] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "b659ff28-5101-4825-84ea-111351c81145-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.830579] env[61985]: INFO nova.compute.manager [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Terminating instance [ 810.831365] env[61985]: DEBUG nova.compute.manager [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 810.831562] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 810.833238] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f59e150-e126-439e-8439-d9697bf59bef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.842350] env[61985]: DEBUG nova.compute.manager [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 810.844702] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 810.845592] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b6ab1b9-ce11-4641-b4bd-a6b2b5de6365 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.853241] env[61985]: DEBUG oslo_vmware.api [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 810.853241] env[61985]: value = "task-935786" [ 810.853241] env[61985]: _type = "Task" [ 810.853241] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.862669] env[61985]: DEBUG oslo_vmware.api [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935786, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.881869] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 810.882599] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 810.882916] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.883249] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 810.883508] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.885210] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 810.885210] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 810.885210] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 810.885210] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 810.885210] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 810.885422] env[61985]: DEBUG nova.virt.hardware [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 810.886044] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ba1273-4611-4a31-92d1-bdfa8d881b6a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.894512] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a542b0b-3c75-4f96-a7e5-53117b1f9ef4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.960177] env[61985]: DEBUG nova.scheduler.client.report [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.018063] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "794cff68-6e26-4607-96f7-eaeb41182551" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.018383] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "794cff68-6e26-4607-96f7-eaeb41182551" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.018570] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "794cff68-6e26-4607-96f7-eaeb41182551-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.018795] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "794cff68-6e26-4607-96f7-eaeb41182551-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.018932] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "794cff68-6e26-4607-96f7-eaeb41182551-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.028584] env[61985]: INFO nova.compute.manager [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Terminating instance [ 811.032099] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "refresh_cache-794cff68-6e26-4607-96f7-eaeb41182551" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.032325] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquired lock "refresh_cache-794cff68-6e26-4607-96f7-eaeb41182551" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.032512] env[61985]: DEBUG nova.network.neutron [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.038982] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935785, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.049604] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52650b8d-b3b4-fcbd-6e67-cbf7a415b702, 'name': SearchDatastore_Task, 'duration_secs': 0.020323} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.050637] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.050983] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.051202] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.051421] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.051534] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.052441] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3124db7e-cc08-4e79-9ada-226ee1ab5f31 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.063368] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.063368] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 811.063687] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8f4bbe7-ace6-4320-8b8c-128fac3dd87d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.070632] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for the task: (returnval){ [ 811.070632] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e4ac1c-4e36-4a0b-4a81-70c6b3988b90" [ 811.070632] env[61985]: _type = "Task" [ 811.070632] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.080908] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e4ac1c-4e36-4a0b-4a81-70c6b3988b90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.181025] env[61985]: DEBUG oslo_vmware.api [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935780, 'name': PowerOnVM_Task, 'duration_secs': 1.914406} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.181025] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 811.181025] env[61985]: DEBUG nova.compute.manager [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.181025] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41501dd1-7243-485c-a7bf-38ce6c4eb983 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.186441] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquiring lock "4aca0ecb-4ae6-4400-accd-d71782b2806d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.186897] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lock "4aca0ecb-4ae6-4400-accd-d71782b2806d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.187399] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquiring lock "4aca0ecb-4ae6-4400-accd-d71782b2806d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.187763] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lock "4aca0ecb-4ae6-4400-accd-d71782b2806d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.188157] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lock "4aca0ecb-4ae6-4400-accd-d71782b2806d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.193461] env[61985]: INFO nova.compute.manager [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Terminating instance [ 811.199043] env[61985]: DEBUG nova.compute.manager [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 811.199394] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 811.200388] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a312e86-d01f-41ee-8996-c1bce4eb67a4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.207772] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 811.208239] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c8ccd73-e7da-4d05-92d7-7dddd80ae496 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.216466] env[61985]: DEBUG oslo_vmware.api [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for the task: (returnval){ [ 811.216466] env[61985]: value = "task-935787" [ 811.216466] env[61985]: _type = "Task" [ 811.216466] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.235573] env[61985]: DEBUG oslo_vmware.api [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935787, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.249434] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.363029] env[61985]: DEBUG nova.network.neutron [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Updated VIF entry in instance network info cache for port 5aa010f5-afab-4ea2-80c4-0d0626458099. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 811.363428] env[61985]: DEBUG nova.network.neutron [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Updating instance_info_cache with network_info: [{"id": "5aa010f5-afab-4ea2-80c4-0d0626458099", "address": "fa:16:3e:67:8d:3f", "network": {"id": "1ccd4f44-07f3-492a-a774-e9dfbee72280", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1344902357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cf6728cd4fa40119f91548ed7a717b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "779b8e65-8b9e-427e-af08-910febd65bfa", "external-id": "nsx-vlan-transportzone-906", "segmentation_id": 906, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa010f5-af", "ovs_interfaceid": "5aa010f5-afab-4ea2-80c4-0d0626458099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.373851] env[61985]: DEBUG oslo_vmware.api [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935786, 'name': PowerOffVM_Task, 'duration_secs': 0.289267} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.374084] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 811.374518] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 811.374518] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-44c63c94-af0e-4aa7-9fc1-18e1e48c33f4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.444243] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 811.444496] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 811.444635] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Deleting the datastore file [datastore2] b659ff28-5101-4825-84ea-111351c81145 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 811.444924] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-26e555bf-1a3b-41ba-97b6-21294d373169 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.453895] env[61985]: DEBUG oslo_vmware.api [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for the task: (returnval){ [ 811.453895] env[61985]: value = "task-935789" [ 811.453895] env[61985]: _type = "Task" [ 811.453895] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.462457] env[61985]: DEBUG oslo_vmware.api [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935789, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.465464] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.655s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.465768] env[61985]: INFO nova.compute.manager [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Migrating [ 811.466062] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.466332] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.468094] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.603s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.539175] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935785, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.583762] env[61985]: DEBUG nova.network.neutron [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 811.592453] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e4ac1c-4e36-4a0b-4a81-70c6b3988b90, 'name': SearchDatastore_Task, 'duration_secs': 0.053599} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.594085] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-824390bd-3773-47d5-8793-d7328fde6ac6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.602798] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for the task: (returnval){ [ 811.602798] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52672b6c-f67e-bc07-b8db-73afd483005e" [ 811.602798] env[61985]: _type = "Task" [ 811.602798] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.620023] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52672b6c-f67e-bc07-b8db-73afd483005e, 'name': SearchDatastore_Task, 'duration_secs': 0.010688} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.620023] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.620023] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] e6a57868-ab6e-45d2-9fa0-94efbb79185b/e6a57868-ab6e-45d2-9fa0-94efbb79185b.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 811.620742] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-497dd089-e3a1-49ec-958e-76ef4fc310f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.629397] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for the task: (returnval){ [ 811.629397] env[61985]: value = "task-935790" [ 811.629397] env[61985]: _type = "Task" [ 811.629397] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.640218] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935790, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.712443] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.735824] env[61985]: DEBUG oslo_vmware.api [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935787, 'name': PowerOffVM_Task, 'duration_secs': 0.190685} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.737443] env[61985]: DEBUG nova.network.neutron [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.742020] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 811.742020] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 811.742020] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bce975b4-75db-42cf-a54f-3de9c99ab3bf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.786032] env[61985]: DEBUG nova.network.neutron [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Updating instance_info_cache with network_info: [{"id": "c529f7f6-f908-4a69-944f-b979f622f99f", "address": "fa:16:3e:20:1b:54", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc529f7f6-f9", "ovs_interfaceid": "c529f7f6-f908-4a69-944f-b979f622f99f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.810838] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 811.811024] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 811.811246] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Deleting the datastore file [datastore1] 4aca0ecb-4ae6-4400-accd-d71782b2806d {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 811.812184] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4bd5d780-1060-4a5a-bfc0-07f6099e5f84 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.820967] env[61985]: DEBUG oslo_vmware.api [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for the task: (returnval){ [ 811.820967] env[61985]: value = "task-935792" [ 811.820967] env[61985]: _type = "Task" [ 811.820967] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.833420] env[61985]: DEBUG oslo_vmware.api [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935792, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.837310] env[61985]: DEBUG nova.network.neutron [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Successfully updated port: f734b929-a7bc-4469-b204-3b5afd314844 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 811.866943] env[61985]: DEBUG oslo_concurrency.lockutils [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] Releasing lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.867482] env[61985]: DEBUG nova.compute.manager [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Received event network-changed-bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 811.867482] env[61985]: DEBUG nova.compute.manager [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Refreshing instance network info cache due to event network-changed-bf071457-a423-45ed-9a6e-22e52c290be8. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 811.867892] env[61985]: DEBUG oslo_concurrency.lockutils [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] Acquiring lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.868110] env[61985]: DEBUG oslo_concurrency.lockutils [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] Acquired lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.868303] env[61985]: DEBUG nova.network.neutron [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Refreshing network info cache for port bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.967747] env[61985]: DEBUG oslo_vmware.api [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Task: {'id': task-935789, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177896} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.968866] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 811.969746] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 811.970352] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 811.970720] env[61985]: INFO nova.compute.manager [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] [instance: b659ff28-5101-4825-84ea-111351c81145] Took 1.14 seconds to destroy the instance on the hypervisor. [ 811.971421] env[61985]: DEBUG oslo.service.loopingcall [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.982802] env[61985]: INFO nova.compute.rpcapi [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 811.983919] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.996735] env[61985]: DEBUG nova.compute.manager [-] [instance: b659ff28-5101-4825-84ea-111351c81145] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 811.996735] env[61985]: DEBUG nova.network.neutron [-] [instance: b659ff28-5101-4825-84ea-111351c81145] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 812.039442] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935785, 'name': ReconfigVM_Task, 'duration_secs': 1.266589} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.040160] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 8f6cd002-b3c2-4276-b195-15d09a143d31/8f6cd002-b3c2-4276-b195-15d09a143d31.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.041604] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e342440-8b11-4d63-9d45-620aecd78f43 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.050907] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 812.050907] env[61985]: value = "task-935793" [ 812.050907] env[61985]: _type = "Task" [ 812.050907] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.065990] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935793, 'name': Rename_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.140742] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935790, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.218047] env[61985]: DEBUG nova.compute.manager [req-53482392-9db3-4dbf-bc13-0908c58d7b55 req-182ff60e-1eef-4e0c-bb53-1c5420e34796 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Received event network-vif-plugged-f734b929-a7bc-4469-b204-3b5afd314844 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 812.218839] env[61985]: DEBUG oslo_concurrency.lockutils [req-53482392-9db3-4dbf-bc13-0908c58d7b55 req-182ff60e-1eef-4e0c-bb53-1c5420e34796 service nova] Acquiring lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.218839] env[61985]: DEBUG oslo_concurrency.lockutils [req-53482392-9db3-4dbf-bc13-0908c58d7b55 req-182ff60e-1eef-4e0c-bb53-1c5420e34796 service nova] Lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.219274] env[61985]: DEBUG oslo_concurrency.lockutils [req-53482392-9db3-4dbf-bc13-0908c58d7b55 req-182ff60e-1eef-4e0c-bb53-1c5420e34796 service nova] Lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.219612] env[61985]: DEBUG nova.compute.manager [req-53482392-9db3-4dbf-bc13-0908c58d7b55 req-182ff60e-1eef-4e0c-bb53-1c5420e34796 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] No waiting events found dispatching network-vif-plugged-f734b929-a7bc-4469-b204-3b5afd314844 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.219845] env[61985]: WARNING nova.compute.manager [req-53482392-9db3-4dbf-bc13-0908c58d7b55 req-182ff60e-1eef-4e0c-bb53-1c5420e34796 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Received unexpected event network-vif-plugged-f734b929-a7bc-4469-b204-3b5afd314844 for instance with vm_state building and task_state spawning. [ 812.241823] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Releasing lock "refresh_cache-794cff68-6e26-4607-96f7-eaeb41182551" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.243394] env[61985]: DEBUG nova.compute.manager [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 812.243394] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 812.244042] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fecfed-b43c-4277-85e2-8b1a8c478809 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.258054] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 812.258054] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d12f601a-b7e1-4f24-ae0a-53395051ffce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.265914] env[61985]: DEBUG oslo_vmware.api [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 812.265914] env[61985]: value = "task-935794" [ 812.265914] env[61985]: _type = "Task" [ 812.265914] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.277990] env[61985]: DEBUG oslo_vmware.api [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935794, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.289674] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "refresh_cache-cfd59b61-cca9-48d5-85e1-1f45d13f1e88" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.346259] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquiring lock "refresh_cache-4a94a6f9-1c86-4628-aa63-341f2c114e2a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.346259] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquired lock "refresh_cache-4a94a6f9-1c86-4628-aa63-341f2c114e2a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.346259] env[61985]: DEBUG nova.network.neutron [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 812.347736] env[61985]: DEBUG oslo_vmware.api [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Task: {'id': task-935792, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.385468} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.350432] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 812.350432] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 812.350432] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 812.350432] env[61985]: INFO nova.compute.manager [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 812.350432] env[61985]: DEBUG oslo.service.loopingcall [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.351108] env[61985]: DEBUG nova.compute.manager [-] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 812.351154] env[61985]: DEBUG nova.network.neutron [-] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 812.418717] env[61985]: INFO nova.network.neutron [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Port bf071457-a423-45ed-9a6e-22e52c290be8 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 812.418891] env[61985]: DEBUG nova.network.neutron [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.502144] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Applying migration context for instance fea0cf39-e851-409f-86f5-31cc128a44dc as it has an incoming, in-progress migration feff90ec-33ae-4a19-ab75-42f6fbb33485. Migration status is pre-migrating {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 812.502740] env[61985]: INFO nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating resource usage from migration feff90ec-33ae-4a19-ab75-42f6fbb33485 [ 812.512602] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.512973] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.513341] env[61985]: DEBUG nova.network.neutron [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 812.531520] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 2322ab72-9841-41fb-9d60-2812baabe108 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.531999] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 794cff68-6e26-4607-96f7-eaeb41182551 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.531999] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 3694e20c-ce37-4097-9991-8a06f38b2734 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.531999] env[61985]: WARNING nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 3815af53-ba3b-4c60-a5fd-2d94102cdd42 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 812.532196] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance f8bc8b71-0317-479d-b2f9-9471757f0774 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.532196] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance b659ff28-5101-4825-84ea-111351c81145 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.532310] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 761d5954-a5ca-4459-a1d6-bfc59b284bf4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.532468] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 4aca0ecb-4ae6-4400-accd-d71782b2806d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.532541] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance cfd59b61-cca9-48d5-85e1-1f45d13f1e88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.532652] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8f6cd002-b3c2-4276-b195-15d09a143d31 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.532874] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance e6a57868-ab6e-45d2-9fa0-94efbb79185b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.533022] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 4a94a6f9-1c86-4628-aa63-341f2c114e2a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 812.565661] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935793, 'name': Rename_Task, 'duration_secs': 0.246241} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.565719] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 812.565992] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26115091-ad04-4ce3-a897-ffe1085a8876 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.576826] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 812.576826] env[61985]: value = "task-935795" [ 812.576826] env[61985]: _type = "Task" [ 812.576826] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.587487] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935795, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.640227] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935790, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524448} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.640541] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] e6a57868-ab6e-45d2-9fa0-94efbb79185b/e6a57868-ab6e-45d2-9fa0-94efbb79185b.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 812.641110] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.641110] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44f41b88-ecc8-48fa-8573-3b95215c4f30 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.648623] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for the task: (returnval){ [ 812.648623] env[61985]: value = "task-935796" [ 812.648623] env[61985]: _type = "Task" [ 812.648623] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.656959] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935796, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.778983] env[61985]: DEBUG oslo_vmware.api [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935794, 'name': PowerOffVM_Task, 'duration_secs': 0.149381} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.778983] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 812.778983] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 812.778983] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46bd67c1-a601-46af-8a8e-070c3a5263b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.802182] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 812.802182] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 812.802182] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Deleting the datastore file [datastore1] 794cff68-6e26-4607-96f7-eaeb41182551 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 812.802412] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f95f50e-c17c-439b-9568-6c86f3143fda {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.809384] env[61985]: DEBUG oslo_vmware.api [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for the task: (returnval){ [ 812.809384] env[61985]: value = "task-935798" [ 812.809384] env[61985]: _type = "Task" [ 812.809384] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.817852] env[61985]: DEBUG oslo_vmware.api [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935798, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.833806] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 812.834079] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-765f92e6-3da0-48ce-8d56-60b48c9efc36 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.840782] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 812.840782] env[61985]: value = "task-935799" [ 812.840782] env[61985]: _type = "Task" [ 812.840782] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.853792] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935799, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.909525] env[61985]: DEBUG nova.network.neutron [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 812.921292] env[61985]: DEBUG oslo_concurrency.lockutils [req-0de276f0-3013-431d-b0e4-c1b8d7482e2f req-87a4f2d2-a090-4f81-b6ef-d6c9da6cb918 service nova] Releasing lock "refresh_cache-b659ff28-5101-4825-84ea-111351c81145" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.946722] env[61985]: DEBUG nova.network.neutron [-] [instance: b659ff28-5101-4825-84ea-111351c81145] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.040039] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 402ef006-2835-4239-a3c4-1c37a7bba9a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 813.088302] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935795, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.106436] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "f8bc8b71-0317-479d-b2f9-9471757f0774" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.107166] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "f8bc8b71-0317-479d-b2f9-9471757f0774" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.108074] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "f8bc8b71-0317-479d-b2f9-9471757f0774-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.108732] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "f8bc8b71-0317-479d-b2f9-9471757f0774-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.114047] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "f8bc8b71-0317-479d-b2f9-9471757f0774-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.004s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.117253] env[61985]: INFO nova.compute.manager [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Terminating instance [ 813.119593] env[61985]: DEBUG nova.compute.manager [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 813.120043] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 813.121821] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084a9c93-531b-4e12-892b-73cc3dcbb492 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.132254] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.132811] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99e66011-d189-4f9d-889c-8d389f7b42a8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.140745] env[61985]: DEBUG oslo_vmware.api [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 813.140745] env[61985]: value = "task-935800" [ 813.140745] env[61985]: _type = "Task" [ 813.140745] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.158025] env[61985]: DEBUG oslo_vmware.api [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935800, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.164250] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935796, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.300473} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.164717] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.165750] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86dd67a-8e3d-45a1-9d1e-54dbdea5e92c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.194714] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] e6a57868-ab6e-45d2-9fa0-94efbb79185b/e6a57868-ab6e-45d2-9fa0-94efbb79185b.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.199282] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b793aec4-b8f9-42ae-87c6-20d2bb367aca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.217471] env[61985]: DEBUG nova.network.neutron [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Updating instance_info_cache with network_info: [{"id": "f734b929-a7bc-4469-b204-3b5afd314844", "address": "fa:16:3e:9d:2c:03", "network": {"id": "6660c6dc-808f-4e2d-aae1-bf49d1809c6d", "bridge": "br-int", "label": "tempest-ServersTestJSON-1808545811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1b149ad11914d118f261481314b5a9d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf734b929-a7", "ovs_interfaceid": "f734b929-a7bc-4469-b204-3b5afd314844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.221928] env[61985]: DEBUG nova.compute.manager [req-c7238d07-e704-4879-ba4e-b0a355b499c5 req-d324f39b-61ce-4fee-bfd3-ec28bf150fd5 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Received event network-vif-deleted-9beae88f-5110-4296-bf01-5b1c423e9017 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 813.222309] env[61985]: INFO nova.compute.manager [req-c7238d07-e704-4879-ba4e-b0a355b499c5 req-d324f39b-61ce-4fee-bfd3-ec28bf150fd5 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Neutron deleted interface 9beae88f-5110-4296-bf01-5b1c423e9017; detaching it from the instance and deleting it from the info cache [ 813.223017] env[61985]: DEBUG nova.network.neutron [req-c7238d07-e704-4879-ba4e-b0a355b499c5 req-d324f39b-61ce-4fee-bfd3-ec28bf150fd5 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.226457] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for the task: (returnval){ [ 813.226457] env[61985]: value = "task-935801" [ 813.226457] env[61985]: _type = "Task" [ 813.226457] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.237198] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935801, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.307473] env[61985]: DEBUG nova.network.neutron [-] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.320480] env[61985]: DEBUG oslo_vmware.api [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Task: {'id': task-935798, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.437178} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.320813] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 813.321415] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 813.321415] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 813.321415] env[61985]: INFO nova.compute.manager [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Took 1.08 seconds to destroy the instance on the hypervisor. [ 813.321657] env[61985]: DEBUG oslo.service.loopingcall [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.322125] env[61985]: DEBUG nova.compute.manager [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 813.322125] env[61985]: DEBUG nova.network.neutron [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 813.347058] env[61985]: DEBUG nova.network.neutron [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.351954] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935799, 'name': PowerOffVM_Task, 'duration_secs': 0.269819} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.352481] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 813.353331] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7159e666-7c17-454d-8f9f-cf23644075c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.377813] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912ac130-2ac3-41dd-905f-2c221873dddc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.416593] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.417283] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d4c7436-ee07-4fca-9342-42836b6e4651 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.426471] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 813.426471] env[61985]: value = "task-935802" [ 813.426471] env[61985]: _type = "Task" [ 813.426471] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.436288] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 813.436512] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 813.436755] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.436904] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.437153] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 813.437364] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a793830d-ef9e-477a-924a-79bc248e311b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.448457] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 813.448752] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 813.449874] env[61985]: INFO nova.compute.manager [-] [instance: b659ff28-5101-4825-84ea-111351c81145] Took 1.45 seconds to deallocate network for instance. [ 813.450135] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ba694c8-b916-438d-bcbc-000346fb1dbe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.463218] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 813.463218] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52680799-5037-c03f-f07d-fdd7ceaf273f" [ 813.463218] env[61985]: _type = "Task" [ 813.463218] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.472511] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52680799-5037-c03f-f07d-fdd7ceaf273f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.544533] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance c4e95f51-9acc-46e8-9921-e0a85fb38598 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 813.588440] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935795, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.599486] env[61985]: DEBUG nova.network.neutron [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance_info_cache with network_info: [{"id": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "address": "fa:16:3e:7b:a1:a8", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadde3d48-eb", "ovs_interfaceid": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.652892] env[61985]: DEBUG oslo_vmware.api [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935800, 'name': PowerOffVM_Task, 'duration_secs': 0.390461} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.652892] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 813.652892] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 813.652892] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77b5a078-679c-426e-bae2-cfcabe299cb8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.722059] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Releasing lock "refresh_cache-4a94a6f9-1c86-4628-aa63-341f2c114e2a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.722473] env[61985]: DEBUG nova.compute.manager [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Instance network_info: |[{"id": "f734b929-a7bc-4469-b204-3b5afd314844", "address": "fa:16:3e:9d:2c:03", "network": {"id": "6660c6dc-808f-4e2d-aae1-bf49d1809c6d", "bridge": "br-int", "label": "tempest-ServersTestJSON-1808545811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1b149ad11914d118f261481314b5a9d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf734b929-a7", "ovs_interfaceid": "f734b929-a7bc-4469-b204-3b5afd314844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 813.723515] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:2c:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '674802e7-b847-4bef-a7a8-f90ac7a3a0a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f734b929-a7bc-4469-b204-3b5afd314844', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.731339] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Creating folder: Project (d1b149ad11914d118f261481314b5a9d). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 813.732585] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05ef27f5-32d9-4ef3-b82d-17dba9d982a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.736465] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f0c1bc6-0a24-4474-a9eb-6c233cf45b02 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.742566] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 813.742777] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 813.742953] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleting the datastore file [datastore2] f8bc8b71-0317-479d-b2f9-9471757f0774 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 813.743999] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65feab9f-d087-4a73-a26f-5a3206f9142f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.753077] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935801, 'name': ReconfigVM_Task, 'duration_secs': 0.4451} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.756675] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Reconfigured VM instance instance-00000031 to attach disk [datastore2] e6a57868-ab6e-45d2-9fa0-94efbb79185b/e6a57868-ab6e-45d2-9fa0-94efbb79185b.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.757664] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Created folder: Project (d1b149ad11914d118f261481314b5a9d) in parent group-v211285. [ 813.757843] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Creating folder: Instances. Parent ref: group-v211338. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 813.758680] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d6bfb78-5884-42bd-b714-a253cfdc5926 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.763538] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58a455f-c120-435c-a65b-9f6983478858 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.774562] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e6204ca-00b9-452f-a568-fd2101eda91e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.779854] env[61985]: DEBUG oslo_vmware.api [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 813.779854] env[61985]: value = "task-935805" [ 813.779854] env[61985]: _type = "Task" [ 813.779854] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.783492] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for the task: (returnval){ [ 813.783492] env[61985]: value = "task-935806" [ 813.783492] env[61985]: _type = "Task" [ 813.783492] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.786698] env[61985]: DEBUG oslo_vmware.api [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935805, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.802016] env[61985]: DEBUG nova.compute.manager [req-c7238d07-e704-4879-ba4e-b0a355b499c5 req-d324f39b-61ce-4fee-bfd3-ec28bf150fd5 service nova] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Detach interface failed, port_id=9beae88f-5110-4296-bf01-5b1c423e9017, reason: Instance 4aca0ecb-4ae6-4400-accd-d71782b2806d could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 813.802016] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Created folder: Instances in parent group-v211338. [ 813.802016] env[61985]: DEBUG oslo.service.loopingcall [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.802016] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 813.802016] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3dd79deb-9505-473b-99ff-b199c05fccf2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.821418] env[61985]: INFO nova.compute.manager [-] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Took 1.47 seconds to deallocate network for instance. [ 813.822064] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935806, 'name': Rename_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.828789] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.828789] env[61985]: value = "task-935808" [ 813.828789] env[61985]: _type = "Task" [ 813.828789] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.837275] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935808, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.852965] env[61985]: DEBUG nova.network.neutron [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.961075] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.974707] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52680799-5037-c03f-f07d-fdd7ceaf273f, 'name': SearchDatastore_Task, 'duration_secs': 0.013042} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.975739] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6959a785-ac57-42ae-a064-7f4dab0df0a5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.982310] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 813.982310] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f433f0-2d23-eb55-5e3e-7e98eb0ad633" [ 813.982310] env[61985]: _type = "Task" [ 813.982310] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.991859] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f433f0-2d23-eb55-5e3e-7e98eb0ad633, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.996784] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.997118] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.048416] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 814.089103] env[61985]: DEBUG oslo_vmware.api [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935795, 'name': PowerOnVM_Task, 'duration_secs': 1.033794} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.089449] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 814.089838] env[61985]: INFO nova.compute.manager [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Took 10.54 seconds to spawn the instance on the hypervisor. [ 814.089989] env[61985]: DEBUG nova.compute.manager [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.090949] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56aec4c3-47ab-463c-b4b1-fa86438eb2a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.102959] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.288689] env[61985]: DEBUG oslo_vmware.api [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935805, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321751} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.292239] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.292416] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 814.292872] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 814.293085] env[61985]: INFO nova.compute.manager [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Took 1.17 seconds to destroy the instance on the hypervisor. [ 814.293328] env[61985]: DEBUG oslo.service.loopingcall [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.293515] env[61985]: DEBUG nova.compute.manager [-] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 814.293611] env[61985]: DEBUG nova.network.neutron [-] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 814.300261] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935806, 'name': Rename_Task, 'duration_secs': 0.257886} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.300324] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 814.300500] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d505acc0-5357-46b4-acee-c2514d2015a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.307323] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for the task: (returnval){ [ 814.307323] env[61985]: value = "task-935809" [ 814.307323] env[61985]: _type = "Task" [ 814.307323] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.318614] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935809, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.329681] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.338009] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935808, 'name': CreateVM_Task, 'duration_secs': 0.355458} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.338180] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 814.338952] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.339126] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.339669] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.339993] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27a9f9a0-dce5-45b9-840f-c29d3e6c7d7d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.344504] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for the task: (returnval){ [ 814.344504] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52764d5b-bc6d-4cfc-9591-fed7d2ede41f" [ 814.344504] env[61985]: _type = "Task" [ 814.344504] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.352217] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52764d5b-bc6d-4cfc-9591-fed7d2ede41f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.356609] env[61985]: INFO nova.compute.manager [-] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Took 1.03 seconds to deallocate network for instance. [ 814.474725] env[61985]: DEBUG nova.compute.manager [req-d0fcfd7c-e8e5-4dd9-8df2-a6d278bc5331 req-a9f6d3c4-b18f-40c9-932e-1868ef53786e service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Received event network-changed-f734b929-a7bc-4469-b204-3b5afd314844 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 814.475148] env[61985]: DEBUG nova.compute.manager [req-d0fcfd7c-e8e5-4dd9-8df2-a6d278bc5331 req-a9f6d3c4-b18f-40c9-932e-1868ef53786e service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Refreshing instance network info cache due to event network-changed-f734b929-a7bc-4469-b204-3b5afd314844. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 814.475424] env[61985]: DEBUG oslo_concurrency.lockutils [req-d0fcfd7c-e8e5-4dd9-8df2-a6d278bc5331 req-a9f6d3c4-b18f-40c9-932e-1868ef53786e service nova] Acquiring lock "refresh_cache-4a94a6f9-1c86-4628-aa63-341f2c114e2a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.475579] env[61985]: DEBUG oslo_concurrency.lockutils [req-d0fcfd7c-e8e5-4dd9-8df2-a6d278bc5331 req-a9f6d3c4-b18f-40c9-932e-1868ef53786e service nova] Acquired lock "refresh_cache-4a94a6f9-1c86-4628-aa63-341f2c114e2a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.475776] env[61985]: DEBUG nova.network.neutron [req-d0fcfd7c-e8e5-4dd9-8df2-a6d278bc5331 req-a9f6d3c4-b18f-40c9-932e-1868ef53786e service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Refreshing network info cache for port f734b929-a7bc-4469-b204-3b5afd314844 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 814.494917] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f433f0-2d23-eb55-5e3e-7e98eb0ad633, 'name': SearchDatastore_Task, 'duration_secs': 0.013552} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.495937] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.496244] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] cfd59b61-cca9-48d5-85e1-1f45d13f1e88/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk. {{(pid=61985) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 814.496519] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf537408-e8ac-4286-9e8e-b06b4d3f3a25 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.507148] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 814.507148] env[61985]: value = "task-935810" [ 814.507148] env[61985]: _type = "Task" [ 814.507148] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.517082] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935810, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.551701] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 814.620920] env[61985]: INFO nova.compute.manager [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Took 30.50 seconds to build instance. [ 814.818316] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935809, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.855973] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52764d5b-bc6d-4cfc-9591-fed7d2ede41f, 'name': SearchDatastore_Task, 'duration_secs': 0.028229} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.856333] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.856578] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.856850] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.857028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.857264] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.857677] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8753f4d-9bae-4cd7-97d6-ea5887ce83b6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.864078] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.867873] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.868147] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 814.868958] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66e67c35-a3fd-4e84-90a6-583cebb661d5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.875496] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for the task: (returnval){ [ 814.875496] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b19cc2-ce41-fa94-312a-acff44a33e6f" [ 814.875496] env[61985]: _type = "Task" [ 814.875496] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.885969] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b19cc2-ce41-fa94-312a-acff44a33e6f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.016638] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935810, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.057188] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 95e3e6d7-9203-47a0-bad7-050eba09c511 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 815.083967] env[61985]: DEBUG nova.network.neutron [-] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.124222] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b4719b07-db03-4ca8-8ad4-6ee58a8ae483 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8f6cd002-b3c2-4276-b195-15d09a143d31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.405s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.215759] env[61985]: DEBUG nova.network.neutron [req-d0fcfd7c-e8e5-4dd9-8df2-a6d278bc5331 req-a9f6d3c4-b18f-40c9-932e-1868ef53786e service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Updated VIF entry in instance network info cache for port f734b929-a7bc-4469-b204-3b5afd314844. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 815.216147] env[61985]: DEBUG nova.network.neutron [req-d0fcfd7c-e8e5-4dd9-8df2-a6d278bc5331 req-a9f6d3c4-b18f-40c9-932e-1868ef53786e service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Updating instance_info_cache with network_info: [{"id": "f734b929-a7bc-4469-b204-3b5afd314844", "address": "fa:16:3e:9d:2c:03", "network": {"id": "6660c6dc-808f-4e2d-aae1-bf49d1809c6d", "bridge": "br-int", "label": "tempest-ServersTestJSON-1808545811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1b149ad11914d118f261481314b5a9d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf734b929-a7", "ovs_interfaceid": "f734b929-a7bc-4469-b204-3b5afd314844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.319043] env[61985]: DEBUG oslo_vmware.api [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935809, 'name': PowerOnVM_Task, 'duration_secs': 0.938745} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.323277] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 815.323534] env[61985]: INFO nova.compute.manager [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Took 9.13 seconds to spawn the instance on the hypervisor. [ 815.323724] env[61985]: DEBUG nova.compute.manager [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 815.324631] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a936f250-950c-4813-8516-d543794ad3aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.389515] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b19cc2-ce41-fa94-312a-acff44a33e6f, 'name': SearchDatastore_Task, 'duration_secs': 0.020935} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.390504] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4877d0ed-f689-4f47-b60b-4e8ab58fa524 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.396176] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for the task: (returnval){ [ 815.396176] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529ce2bf-f661-50a8-f5b1-66b940dd17f9" [ 815.396176] env[61985]: _type = "Task" [ 815.396176] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.403982] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529ce2bf-f661-50a8-f5b1-66b940dd17f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.522828] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935810, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523425} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.523159] env[61985]: INFO nova.virt.vmwareapi.ds_util [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] cfd59b61-cca9-48d5-85e1-1f45d13f1e88/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk. [ 815.523977] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9df96ea-3c1a-4edd-9408-d50ef986002d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.554504] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] cfd59b61-cca9-48d5-85e1-1f45d13f1e88/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.554504] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ae848de-59c5-4c28-8040-1cae89ae6a57 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.568642] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance e82d26b1-5502-4fd2-89c5-ffb0c1557c79 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 815.576276] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 815.576276] env[61985]: value = "task-935811" [ 815.576276] env[61985]: _type = "Task" [ 815.576276] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.586026] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935811, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.586466] env[61985]: INFO nova.compute.manager [-] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Took 1.29 seconds to deallocate network for instance. [ 815.626969] env[61985]: DEBUG nova.compute.manager [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 815.632031] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061dd723-7d89-431f-9518-2994a486b593 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.657748] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance 'fea0cf39-e851-409f-86f5-31cc128a44dc' progress to 0 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 815.718693] env[61985]: DEBUG oslo_concurrency.lockutils [req-d0fcfd7c-e8e5-4dd9-8df2-a6d278bc5331 req-a9f6d3c4-b18f-40c9-932e-1868ef53786e service nova] Releasing lock "refresh_cache-4a94a6f9-1c86-4628-aa63-341f2c114e2a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.718928] env[61985]: DEBUG nova.compute.manager [req-d0fcfd7c-e8e5-4dd9-8df2-a6d278bc5331 req-a9f6d3c4-b18f-40c9-932e-1868ef53786e service nova] [instance: b659ff28-5101-4825-84ea-111351c81145] Received event network-vif-deleted-bf071457-a423-45ed-9a6e-22e52c290be8 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 815.843671] env[61985]: INFO nova.compute.manager [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Took 26.73 seconds to build instance. [ 815.911046] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529ce2bf-f661-50a8-f5b1-66b940dd17f9, 'name': SearchDatastore_Task, 'duration_secs': 0.091107} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.911046] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.911046] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 4a94a6f9-1c86-4628-aa63-341f2c114e2a/4a94a6f9-1c86-4628-aa63-341f2c114e2a.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 815.911046] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d67593e-e31f-417f-b28d-314fc81e3ea1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.919116] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for the task: (returnval){ [ 815.919116] env[61985]: value = "task-935812" [ 815.919116] env[61985]: _type = "Task" [ 815.919116] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.927721] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935812, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.072116] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance f8924f18-7232-4be6-84bf-1ba05b855ffe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 816.087393] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935811, 'name': ReconfigVM_Task, 'duration_secs': 0.332892} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.087701] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Reconfigured VM instance instance-0000002f to attach disk [datastore1] cfd59b61-cca9-48d5-85e1-1f45d13f1e88/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.088576] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4abb9ab-f48c-483f-88b8-9bd33c26451e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.112347] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.118029] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0eaa567-fac5-42f2-954c-61fa68f3067e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.133771] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 816.133771] env[61985]: value = "task-935813" [ 816.133771] env[61985]: _type = "Task" [ 816.133771] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.143155] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935813, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.161414] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.164147] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 816.164426] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbfbaf87-6e6d-42fc-bf8f-e4302e62cb3d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.171530] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 816.171530] env[61985]: value = "task-935814" [ 816.171530] env[61985]: _type = "Task" [ 816.171530] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.183498] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.200377] env[61985]: DEBUG oslo_concurrency.lockutils [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "8f6cd002-b3c2-4276-b195-15d09a143d31" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.200679] env[61985]: DEBUG oslo_concurrency.lockutils [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8f6cd002-b3c2-4276-b195-15d09a143d31" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.200897] env[61985]: DEBUG oslo_concurrency.lockutils [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "8f6cd002-b3c2-4276-b195-15d09a143d31-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.201097] env[61985]: DEBUG oslo_concurrency.lockutils [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8f6cd002-b3c2-4276-b195-15d09a143d31-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.201271] env[61985]: DEBUG oslo_concurrency.lockutils [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8f6cd002-b3c2-4276-b195-15d09a143d31-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.203980] env[61985]: INFO nova.compute.manager [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Terminating instance [ 816.205909] env[61985]: DEBUG nova.compute.manager [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 816.206139] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 816.207193] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a60f23-42ff-4ef6-84e4-2c5e4d75dd11 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.214517] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 816.215586] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49e63327-524c-4999-9af4-84515ebb157c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.222101] env[61985]: DEBUG oslo_vmware.api [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 816.222101] env[61985]: value = "task-935815" [ 816.222101] env[61985]: _type = "Task" [ 816.222101] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.231478] env[61985]: DEBUG oslo_vmware.api [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935815, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.346814] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f83d8284-e14b-4773-b504-eda6194530fe tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.242s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.432964] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935812, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.576292] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 1f284789-1e7b-4e9f-9670-34e8e25cd797 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 816.645547] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935813, 'name': ReconfigVM_Task, 'duration_secs': 0.150433} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.645547] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 816.645547] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bdc6b1e4-83ab-41cb-a14c-cf973d7f12f0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.652452] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 816.652452] env[61985]: value = "task-935816" [ 816.652452] env[61985]: _type = "Task" [ 816.652452] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.661282] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935816, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.681131] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935814, 'name': PowerOffVM_Task, 'duration_secs': 0.201079} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.681131] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 816.681311] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance 'fea0cf39-e851-409f-86f5-31cc128a44dc' progress to 17 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 816.689286] env[61985]: DEBUG nova.compute.manager [req-e8b1db92-0cdb-49d8-b8ce-36441d14ea8c req-21179ee5-7ad7-4699-abe8-b0ee85556c38 service nova] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Received event network-vif-deleted-787b8b7e-dea3-41b7-9349-4f87afe7b88d {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 816.734926] env[61985]: DEBUG oslo_vmware.api [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935815, 'name': PowerOffVM_Task, 'duration_secs': 0.377057} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.735678] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 816.736264] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 816.736973] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07681dde-04d8-4ca7-859a-7d4feb0081e0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.852791] env[61985]: DEBUG nova.compute.manager [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.883845] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 816.884467] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 816.884815] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleting the datastore file [datastore2] 8f6cd002-b3c2-4276-b195-15d09a143d31 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.885213] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30ec495d-c2a8-41f1-a5da-f33ac42a1cd4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.892465] env[61985]: DEBUG oslo_vmware.api [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 816.892465] env[61985]: value = "task-935818" [ 816.892465] env[61985]: _type = "Task" [ 816.892465] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.901746] env[61985]: DEBUG oslo_vmware.api [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.931900] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935812, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.724938} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.932574] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 4a94a6f9-1c86-4628-aa63-341f2c114e2a/4a94a6f9-1c86-4628-aa63-341f2c114e2a.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 816.932988] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.933511] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6e704c35-284d-46af-b82f-ef6bf03ad7a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.940791] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for the task: (returnval){ [ 816.940791] env[61985]: value = "task-935819" [ 816.940791] env[61985]: _type = "Task" [ 816.940791] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.949311] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935819, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.081630] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 17bdf8ec-13d2-459c-bc8e-db6a274fc27e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 817.163248] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935816, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.188848] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.188848] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.188848] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.188848] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.189169] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.192395] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.192395] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.192395] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.192395] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.192395] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.192668] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.196218] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-554f222d-8ff5-4345-ae16-2b00e8e2f412 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.216135] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 817.216135] env[61985]: value = "task-935820" [ 817.216135] env[61985]: _type = "Task" [ 817.216135] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.225158] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935820, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.375438] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.406636] env[61985]: DEBUG oslo_vmware.api [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-935818, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138715} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.406636] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.406636] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 817.406636] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 817.406636] env[61985]: INFO nova.compute.manager [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Took 1.20 seconds to destroy the instance on the hypervisor. [ 817.406958] env[61985]: DEBUG oslo.service.loopingcall [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.407166] env[61985]: DEBUG nova.compute.manager [-] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.407295] env[61985]: DEBUG nova.network.neutron [-] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 817.459060] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935819, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068204} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.459200] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.459984] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8616dd7-258c-44d9-8848-4a74a77abd19 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.488929] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 4a94a6f9-1c86-4628-aa63-341f2c114e2a/4a94a6f9-1c86-4628-aa63-341f2c114e2a.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.489940] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a8ba469-391c-49a9-bb9e-b8f8c196a3ec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.513121] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for the task: (returnval){ [ 817.513121] env[61985]: value = "task-935821" [ 817.513121] env[61985]: _type = "Task" [ 817.513121] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.521748] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935821, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.583767] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 59e742d9-c51a-47dd-9ef2-2f070eae94c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 817.663508] env[61985]: DEBUG oslo_vmware.api [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935816, 'name': PowerOnVM_Task, 'duration_secs': 0.915851} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.663825] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 817.667608] env[61985]: DEBUG nova.compute.manager [None req-1299a979-1103-473b-8436-c70e00cd8d49 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.669080] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82741ec-632a-4ca8-9873-b3be8ff7f703 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.730647] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935820, 'name': ReconfigVM_Task, 'duration_secs': 0.432826} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.731025] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance 'fea0cf39-e851-409f-86f5-31cc128a44dc' progress to 33 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 818.029068] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935821, 'name': ReconfigVM_Task, 'duration_secs': 0.497479} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.029068] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 4a94a6f9-1c86-4628-aa63-341f2c114e2a/4a94a6f9-1c86-4628-aa63-341f2c114e2a.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 818.029068] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8958f9fb-32bb-41c6-ab31-18e36ccca9c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.035982] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for the task: (returnval){ [ 818.035982] env[61985]: value = "task-935822" [ 818.035982] env[61985]: _type = "Task" [ 818.035982] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.044803] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935822, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.088070] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance ce757dda-f58a-47b3-b319-e6b01f05b20b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 818.239612] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.239612] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.239612] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.239612] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.239993] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.239993] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.240202] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.241044] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.241044] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.241044] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.241044] env[61985]: DEBUG nova.virt.hardware [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.246375] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Reconfiguring VM instance instance-00000027 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 818.246493] env[61985]: DEBUG nova.network.neutron [-] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.247807] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d972be8d-762a-4318-978c-aaf4557c1187 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.269692] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 818.269692] env[61985]: value = "task-935823" [ 818.269692] env[61985]: _type = "Task" [ 818.269692] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.280421] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935823, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.536368] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquiring lock "518635bf-73ee-404b-ae6a-dc4ee23009d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.536660] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lock "518635bf-73ee-404b-ae6a-dc4ee23009d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.547414] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935822, 'name': Rename_Task, 'duration_secs': 0.353895} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.547980] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 818.548472] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-513b5b6e-91b6-4d0d-bd41-aafe9c73dcc7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.558081] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for the task: (returnval){ [ 818.558081] env[61985]: value = "task-935824" [ 818.558081] env[61985]: _type = "Task" [ 818.558081] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.568120] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935824, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.592053] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 5b3a8ba3-48cb-472a-ab45-4e51af68812c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 818.592697] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Migration feff90ec-33ae-4a19-ab75-42f6fbb33485 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 818.592697] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance fea0cf39-e851-409f-86f5-31cc128a44dc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 818.765195] env[61985]: INFO nova.compute.manager [-] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Took 1.36 seconds to deallocate network for instance. [ 818.783605] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935823, 'name': ReconfigVM_Task, 'duration_secs': 0.243498} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.784097] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Reconfigured VM instance instance-00000027 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 818.784741] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c303c8b-3c93-4001-8008-31a74f8eb55c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.563466] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance c64056d8-0c22-48db-bf4a-a60f468bdad0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 819.563737] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 819.563887] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 819.566950] env[61985]: DEBUG oslo_concurrency.lockutils [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.575449] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] fea0cf39-e851-409f-86f5-31cc128a44dc/fea0cf39-e851-409f-86f5-31cc128a44dc.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.576929] env[61985]: DEBUG nova.compute.manager [req-68bff5fe-9c99-433a-bc5c-c8b25727eeef req-f4f65d0f-3477-4231-823c-846dd8ac8941 service nova] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Received event network-vif-deleted-b6ffda80-c267-47cb-92e7-6a930fbf0c27 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 819.577888] env[61985]: DEBUG nova.compute.manager [req-ddea747f-59ff-49a0-959f-3b478f2446d0 req-bd285486-33db-4cf3-a2bb-a01084384a7f service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Received event network-changed-5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 819.578131] env[61985]: DEBUG nova.compute.manager [req-ddea747f-59ff-49a0-959f-3b478f2446d0 req-bd285486-33db-4cf3-a2bb-a01084384a7f service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Refreshing instance network info cache due to event network-changed-5aa010f5-afab-4ea2-80c4-0d0626458099. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 819.578328] env[61985]: DEBUG oslo_concurrency.lockutils [req-ddea747f-59ff-49a0-959f-3b478f2446d0 req-bd285486-33db-4cf3-a2bb-a01084384a7f service nova] Acquiring lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.578527] env[61985]: DEBUG oslo_concurrency.lockutils [req-ddea747f-59ff-49a0-959f-3b478f2446d0 req-bd285486-33db-4cf3-a2bb-a01084384a7f service nova] Acquired lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.578603] env[61985]: DEBUG nova.network.neutron [req-ddea747f-59ff-49a0-959f-3b478f2446d0 req-bd285486-33db-4cf3-a2bb-a01084384a7f service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Refreshing network info cache for port 5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.583045] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-974d2ff1-b5d8-4224-9c1f-80c35c7c80d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.603667] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935824, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.605183] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 819.605183] env[61985]: value = "task-935825" [ 819.605183] env[61985]: _type = "Task" [ 819.605183] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.617263] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935825, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.051026] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219e4686-1925-4a63-8f04-82e76b88ab6b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.058686] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5cbe35d-d90a-42ba-b3c1-2db93f670802 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.092566] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0e8725-5844-4bc5-a8f9-3c4e153b1c7a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.100680] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935824, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.108344] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4411e4d9-af1c-48f0-8c3d-a9fd83869b4e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.128047] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.132140] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935825, 'name': ReconfigVM_Task, 'duration_secs': 0.361413} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.132925] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Reconfigured VM instance instance-00000027 to attach disk [datastore2] fea0cf39-e851-409f-86f5-31cc128a44dc/fea0cf39-e851-409f-86f5-31cc128a44dc.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.132925] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance 'fea0cf39-e851-409f-86f5-31cc128a44dc' progress to 50 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 820.345340] env[61985]: DEBUG nova.network.neutron [req-ddea747f-59ff-49a0-959f-3b478f2446d0 req-bd285486-33db-4cf3-a2bb-a01084384a7f service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Updated VIF entry in instance network info cache for port 5aa010f5-afab-4ea2-80c4-0d0626458099. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 820.345814] env[61985]: DEBUG nova.network.neutron [req-ddea747f-59ff-49a0-959f-3b478f2446d0 req-bd285486-33db-4cf3-a2bb-a01084384a7f service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Updating instance_info_cache with network_info: [{"id": "5aa010f5-afab-4ea2-80c4-0d0626458099", "address": "fa:16:3e:67:8d:3f", "network": {"id": "1ccd4f44-07f3-492a-a774-e9dfbee72280", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1344902357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cf6728cd4fa40119f91548ed7a717b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "779b8e65-8b9e-427e-af08-910febd65bfa", "external-id": "nsx-vlan-transportzone-906", "segmentation_id": 906, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa010f5-af", "ovs_interfaceid": "5aa010f5-afab-4ea2-80c4-0d0626458099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.573390] env[61985]: DEBUG oslo_vmware.api [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935824, 'name': PowerOnVM_Task, 'duration_secs': 1.915378} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.573621] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 820.573836] env[61985]: INFO nova.compute.manager [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Took 9.73 seconds to spawn the instance on the hypervisor. [ 820.574138] env[61985]: DEBUG nova.compute.manager [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.574905] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dba4457-823d-4ecd-a424-ea4698da3fd3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.634361] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.643110] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b074a0-9757-4189-bbad-3ac2443c30b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.660598] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d804f069-497f-441e-88f5-a8487298fd92 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.680566] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance 'fea0cf39-e851-409f-86f5-31cc128a44dc' progress to 67 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 820.848704] env[61985]: DEBUG oslo_concurrency.lockutils [req-ddea747f-59ff-49a0-959f-3b478f2446d0 req-bd285486-33db-4cf3-a2bb-a01084384a7f service nova] Releasing lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.093971] env[61985]: INFO nova.compute.manager [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Took 31.60 seconds to build instance. [ 821.143236] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 821.143529] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.676s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.143814] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.136s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.145449] env[61985]: INFO nova.compute.claims [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.232779] env[61985]: DEBUG nova.network.neutron [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Port adde3d48-ebd4-450d-920a-fb93eeff4a8b binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 821.595961] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89847fcb-6ecf-4220-a2af-600a773c4d8c tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.013s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.690361] env[61985]: DEBUG nova.compute.manager [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.691502] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0729fd9-a6f9-4342-bf36-3660fc56f834 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.923715] env[61985]: INFO nova.compute.manager [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Rescuing [ 821.923993] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.924163] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.924330] env[61985]: DEBUG nova.network.neutron [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 822.099782] env[61985]: DEBUG nova.compute.manager [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 822.202811] env[61985]: INFO nova.compute.manager [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] instance snapshotting [ 822.203411] env[61985]: DEBUG nova.objects.instance [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'flavor' on Instance uuid 2322ab72-9841-41fb-9d60-2812baabe108 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 822.258549] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "fea0cf39-e851-409f-86f5-31cc128a44dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.258801] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.258977] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.572011] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7e01df-4c4d-41c2-aa2a-454dc5cecb95 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.580542] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63d0235-6230-4f9a-b828-d0291328665a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.620962] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07a66ca-0542-4371-90e4-ccc9fffdbd67 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.637159] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375c36e9-8bc9-4c84-a45c-2875c88b1248 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.641522] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.654919] env[61985]: DEBUG nova.compute.provider_tree [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.669311] env[61985]: DEBUG nova.network.neutron [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Updating instance_info_cache with network_info: [{"id": "6ea4b2df-e4c4-48a4-9af2-fe4606759004", "address": "fa:16:3e:b3:50:b6", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ea4b2df-e4", "ovs_interfaceid": "6ea4b2df-e4c4-48a4-9af2-fe4606759004", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.711115] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799acf25-07ea-4888-b569-438f3a357b49 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.731242] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff3b413-0759-4975-b6b9-fb7873161439 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.161859] env[61985]: DEBUG nova.scheduler.client.report [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.173549] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.243022] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 823.243022] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5e4ab8a0-bcfd-4343-bf6b-c7020eb6128f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.255121] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 823.255121] env[61985]: value = "task-935826" [ 823.255121] env[61985]: _type = "Task" [ 823.255121] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.272525] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935826, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.331560] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.331560] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.331560] env[61985]: DEBUG nova.network.neutron [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 823.670376] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.670376] env[61985]: DEBUG nova.compute.manager [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 823.671792] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.445s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.673197] env[61985]: INFO nova.compute.claims [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.713171] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 823.713532] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e11fbba-dbc6-421b-b273-7d6d94f78f64 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.720244] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 823.720244] env[61985]: value = "task-935827" [ 823.720244] env[61985]: _type = "Task" [ 823.720244] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.730072] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935827, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.755694] env[61985]: DEBUG nova.compute.manager [req-52783e45-82a6-4bc0-8bad-482dd98c413d req-50be091d-f12b-4764-9d35-c16b0d7c0e47 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Received event network-changed-f734b929-a7bc-4469-b204-3b5afd314844 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 823.755863] env[61985]: DEBUG nova.compute.manager [req-52783e45-82a6-4bc0-8bad-482dd98c413d req-50be091d-f12b-4764-9d35-c16b0d7c0e47 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Refreshing instance network info cache due to event network-changed-f734b929-a7bc-4469-b204-3b5afd314844. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 823.756084] env[61985]: DEBUG oslo_concurrency.lockutils [req-52783e45-82a6-4bc0-8bad-482dd98c413d req-50be091d-f12b-4764-9d35-c16b0d7c0e47 service nova] Acquiring lock "refresh_cache-4a94a6f9-1c86-4628-aa63-341f2c114e2a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.756223] env[61985]: DEBUG oslo_concurrency.lockutils [req-52783e45-82a6-4bc0-8bad-482dd98c413d req-50be091d-f12b-4764-9d35-c16b0d7c0e47 service nova] Acquired lock "refresh_cache-4a94a6f9-1c86-4628-aa63-341f2c114e2a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.756379] env[61985]: DEBUG nova.network.neutron [req-52783e45-82a6-4bc0-8bad-482dd98c413d req-50be091d-f12b-4764-9d35-c16b0d7c0e47 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Refreshing network info cache for port f734b929-a7bc-4469-b204-3b5afd314844 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 823.768306] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935826, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.065694] env[61985]: DEBUG nova.network.neutron [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance_info_cache with network_info: [{"id": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "address": "fa:16:3e:7b:a1:a8", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadde3d48-eb", "ovs_interfaceid": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.182646] env[61985]: DEBUG nova.compute.utils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.188399] env[61985]: DEBUG nova.compute.manager [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.188399] env[61985]: DEBUG nova.network.neutron [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 824.231793] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935827, 'name': PowerOffVM_Task, 'duration_secs': 0.245101} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.232485] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 824.233937] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca07be70-f9b8-4296-8984-4baf3c86e4a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.254995] env[61985]: DEBUG nova.policy [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f0993b4a1b8a4b4daf71c69db5993242', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a25034654f4e411d9d2c895e1412de74', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.258044] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61870bba-f398-4b98-9900-55910e481b64 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.276179] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935826, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.303601] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 824.303905] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-386aaa5d-392a-446e-b2f8-9d35be6b06cd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.314050] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 824.314050] env[61985]: value = "task-935828" [ 824.314050] env[61985]: _type = "Task" [ 824.314050] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.326590] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 824.326805] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.327059] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.327211] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.327385] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.327641] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-90030e12-2b98-4912-aa89-bb02afa7dbea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.335889] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.336092] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 824.336824] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4fbb667-4590-40c6-a03c-d07f763a5fc3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.342334] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 824.342334] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5202d633-f897-a43f-2e23-e32ab88156de" [ 824.342334] env[61985]: _type = "Task" [ 824.342334] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.350450] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5202d633-f897-a43f-2e23-e32ab88156de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.569123] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.615300] env[61985]: DEBUG nova.network.neutron [req-52783e45-82a6-4bc0-8bad-482dd98c413d req-50be091d-f12b-4764-9d35-c16b0d7c0e47 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Updated VIF entry in instance network info cache for port f734b929-a7bc-4469-b204-3b5afd314844. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 824.615710] env[61985]: DEBUG nova.network.neutron [req-52783e45-82a6-4bc0-8bad-482dd98c413d req-50be091d-f12b-4764-9d35-c16b0d7c0e47 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Updating instance_info_cache with network_info: [{"id": "f734b929-a7bc-4469-b204-3b5afd314844", "address": "fa:16:3e:9d:2c:03", "network": {"id": "6660c6dc-808f-4e2d-aae1-bf49d1809c6d", "bridge": "br-int", "label": "tempest-ServersTestJSON-1808545811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1b149ad11914d118f261481314b5a9d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf734b929-a7", "ovs_interfaceid": "f734b929-a7bc-4469-b204-3b5afd314844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.635155] env[61985]: DEBUG nova.network.neutron [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Successfully created port: c32c52fa-48c3-48d7-87e9-0e9d693d6959 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 824.689659] env[61985]: DEBUG nova.compute.manager [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 824.776111] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935826, 'name': CreateSnapshot_Task, 'duration_secs': 1.227806} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.777028] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 824.777241] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd692f26-e8da-4b35-a1e0-0affa269f51f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.862926] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5202d633-f897-a43f-2e23-e32ab88156de, 'name': SearchDatastore_Task, 'duration_secs': 0.010868} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.863913] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-917b8ca3-08e2-49e4-ae17-0c381653e85a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.872229] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 824.872229] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52843b23-1305-9db1-0420-ed7ad8e7542d" [ 824.872229] env[61985]: _type = "Task" [ 824.872229] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.887697] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52843b23-1305-9db1-0420-ed7ad8e7542d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.095502] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724396ba-5950-48ed-8f9e-1c8ee418bdb9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.117233] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d34d80c-d0ca-4b29-9f0b-6d81d4658229 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.120582] env[61985]: DEBUG oslo_concurrency.lockutils [req-52783e45-82a6-4bc0-8bad-482dd98c413d req-50be091d-f12b-4764-9d35-c16b0d7c0e47 service nova] Releasing lock "refresh_cache-4a94a6f9-1c86-4628-aa63-341f2c114e2a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.125689] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance 'fea0cf39-e851-409f-86f5-31cc128a44dc' progress to 83 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 825.229901] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c084dc6b-4261-44d7-9e84-22b454056b60 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.240212] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d05733d-f9df-4a5c-ba0d-2bee248bf7f0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.268947] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1302edd5-c91a-479c-96f1-d05dd84e250b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.276855] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb04737a-1b0a-49fc-915d-0b6264dccb6c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.300566] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 825.301138] env[61985]: DEBUG nova.compute.provider_tree [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.302961] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cde95c8d-792d-4072-98eb-1760d0e37312 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.312346] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 825.312346] env[61985]: value = "task-935829" [ 825.312346] env[61985]: _type = "Task" [ 825.312346] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.320932] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935829, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.365512] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.384489] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52843b23-1305-9db1-0420-ed7ad8e7542d, 'name': SearchDatastore_Task, 'duration_secs': 0.014972} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.384760] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.385034] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 761d5954-a5ca-4459-a1d6-bfc59b284bf4/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk. {{(pid=61985) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 825.385306] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ab9e0a1-b7fa-49eb-b32e-58b14f30970d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.392854] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 825.392854] env[61985]: value = "task-935830" [ 825.392854] env[61985]: _type = "Task" [ 825.392854] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.405027] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.431250] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquiring lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.431655] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.432750] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquiring lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.433011] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.433205] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.435851] env[61985]: INFO nova.compute.manager [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Terminating instance [ 825.437999] env[61985]: DEBUG nova.compute.manager [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 825.438207] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 825.439102] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30dd97be-42e0-4dd1-8591-1ba4e841c914 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.447059] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.447585] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ba4bb25-d7ea-49d3-8f20-3db87fa5fc08 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.454609] env[61985]: DEBUG oslo_vmware.api [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for the task: (returnval){ [ 825.454609] env[61985]: value = "task-935831" [ 825.454609] env[61985]: _type = "Task" [ 825.454609] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.463245] env[61985]: DEBUG oslo_vmware.api [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.632167] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 825.632485] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07c77568-4c26-4cf7-b828-e961de8b1dba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.640205] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 825.640205] env[61985]: value = "task-935832" [ 825.640205] env[61985]: _type = "Task" [ 825.640205] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.649754] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935832, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.705438] env[61985]: DEBUG nova.compute.manager [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 825.731498] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.731761] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.731923] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.732217] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.732386] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.732538] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.732756] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.732917] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.733110] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.733280] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.733453] env[61985]: DEBUG nova.virt.hardware [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.734429] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e29c13-5c8e-49e4-b64f-81fee48aef13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.742881] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4704c9-d599-461a-841b-e8b02ec9da77 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.789541] env[61985]: DEBUG nova.compute.manager [req-46670c1b-2981-4d1f-bd07-2adb220582dc req-fd665f4e-ccb1-45db-b657-da8f612a2605 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Received event network-changed-5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 825.789541] env[61985]: DEBUG nova.compute.manager [req-46670c1b-2981-4d1f-bd07-2adb220582dc req-fd665f4e-ccb1-45db-b657-da8f612a2605 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Refreshing instance network info cache due to event network-changed-5aa010f5-afab-4ea2-80c4-0d0626458099. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 825.789541] env[61985]: DEBUG oslo_concurrency.lockutils [req-46670c1b-2981-4d1f-bd07-2adb220582dc req-fd665f4e-ccb1-45db-b657-da8f612a2605 service nova] Acquiring lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.789541] env[61985]: DEBUG oslo_concurrency.lockutils [req-46670c1b-2981-4d1f-bd07-2adb220582dc req-fd665f4e-ccb1-45db-b657-da8f612a2605 service nova] Acquired lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.789541] env[61985]: DEBUG nova.network.neutron [req-46670c1b-2981-4d1f-bd07-2adb220582dc req-fd665f4e-ccb1-45db-b657-da8f612a2605 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Refreshing network info cache for port 5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 825.808586] env[61985]: DEBUG nova.scheduler.client.report [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.825546] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935829, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.874035] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.907034] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935830, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.965571] env[61985]: DEBUG oslo_vmware.api [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935831, 'name': PowerOffVM_Task, 'duration_secs': 0.208554} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.965874] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.966125] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 825.966446] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b09cc9d1-07d4-403b-83dd-87eb5eaa4e0b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.035979] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 826.036242] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 826.036428] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Deleting the datastore file [datastore2] e6a57868-ab6e-45d2-9fa0-94efbb79185b {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 826.036706] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74132748-01fb-4829-90fc-625d8e0c300a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.044645] env[61985]: DEBUG oslo_vmware.api [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for the task: (returnval){ [ 826.044645] env[61985]: value = "task-935834" [ 826.044645] env[61985]: _type = "Task" [ 826.044645] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.053253] env[61985]: DEBUG oslo_vmware.api [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935834, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.151144] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935832, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.314995] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.643s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.315985] env[61985]: DEBUG nova.compute.manager [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 826.319076] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.026s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.319276] env[61985]: DEBUG nova.objects.instance [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61985) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 826.332159] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935829, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.404915] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935830, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595794} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.407404] env[61985]: INFO nova.virt.vmwareapi.ds_util [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 761d5954-a5ca-4459-a1d6-bfc59b284bf4/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk. [ 826.408275] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9ac946-1149-457e-9bbb-3a20f012e226 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.433834] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 761d5954-a5ca-4459-a1d6-bfc59b284bf4/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.436563] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f862aaa-cf81-484e-9a20-6af021321285 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.454859] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 826.454859] env[61985]: value = "task-935835" [ 826.454859] env[61985]: _type = "Task" [ 826.454859] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.462724] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935835, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.557081] env[61985]: DEBUG oslo_vmware.api [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Task: {'id': task-935834, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.362204} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.557538] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.557774] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 826.557985] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 826.558220] env[61985]: INFO nova.compute.manager [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 826.558551] env[61985]: DEBUG oslo.service.loopingcall [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.558767] env[61985]: DEBUG nova.compute.manager [-] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.558873] env[61985]: DEBUG nova.network.neutron [-] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 826.650879] env[61985]: DEBUG oslo_vmware.api [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-935832, 'name': PowerOnVM_Task, 'duration_secs': 0.96782} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.651164] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 826.651345] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18385ae-b767-411a-b519-c50e6bdea8e8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance 'fea0cf39-e851-409f-86f5-31cc128a44dc' progress to 100 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 826.676077] env[61985]: DEBUG nova.network.neutron [req-46670c1b-2981-4d1f-bd07-2adb220582dc req-fd665f4e-ccb1-45db-b657-da8f612a2605 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Updated VIF entry in instance network info cache for port 5aa010f5-afab-4ea2-80c4-0d0626458099. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 826.676553] env[61985]: DEBUG nova.network.neutron [req-46670c1b-2981-4d1f-bd07-2adb220582dc req-fd665f4e-ccb1-45db-b657-da8f612a2605 service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Updating instance_info_cache with network_info: [{"id": "5aa010f5-afab-4ea2-80c4-0d0626458099", "address": "fa:16:3e:67:8d:3f", "network": {"id": "1ccd4f44-07f3-492a-a774-e9dfbee72280", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1344902357-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8cf6728cd4fa40119f91548ed7a717b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "779b8e65-8b9e-427e-af08-910febd65bfa", "external-id": "nsx-vlan-transportzone-906", "segmentation_id": 906, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5aa010f5-af", "ovs_interfaceid": "5aa010f5-afab-4ea2-80c4-0d0626458099", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.832445] env[61985]: DEBUG nova.compute.utils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.834656] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935829, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.835243] env[61985]: DEBUG nova.compute.manager [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.835711] env[61985]: DEBUG nova.network.neutron [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 826.925269] env[61985]: DEBUG nova.policy [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da7c7d5d10b34b4ebdede5ff3cfd8b01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e66b7a6e7354b06a77295eadf4e5e0f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.964968] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.162273] env[61985]: DEBUG nova.compute.manager [req-bf7f7414-d59f-4e5f-91b2-0c6a785b415e req-0965f6b7-e45e-4552-ae06-e6dcd6714215 service nova] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Received event network-vif-plugged-c32c52fa-48c3-48d7-87e9-0e9d693d6959 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 827.163111] env[61985]: DEBUG oslo_concurrency.lockutils [req-bf7f7414-d59f-4e5f-91b2-0c6a785b415e req-0965f6b7-e45e-4552-ae06-e6dcd6714215 service nova] Acquiring lock "402ef006-2835-4239-a3c4-1c37a7bba9a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.163470] env[61985]: DEBUG oslo_concurrency.lockutils [req-bf7f7414-d59f-4e5f-91b2-0c6a785b415e req-0965f6b7-e45e-4552-ae06-e6dcd6714215 service nova] Lock "402ef006-2835-4239-a3c4-1c37a7bba9a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.163797] env[61985]: DEBUG oslo_concurrency.lockutils [req-bf7f7414-d59f-4e5f-91b2-0c6a785b415e req-0965f6b7-e45e-4552-ae06-e6dcd6714215 service nova] Lock "402ef006-2835-4239-a3c4-1c37a7bba9a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.164200] env[61985]: DEBUG nova.compute.manager [req-bf7f7414-d59f-4e5f-91b2-0c6a785b415e req-0965f6b7-e45e-4552-ae06-e6dcd6714215 service nova] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] No waiting events found dispatching network-vif-plugged-c32c52fa-48c3-48d7-87e9-0e9d693d6959 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 827.164546] env[61985]: WARNING nova.compute.manager [req-bf7f7414-d59f-4e5f-91b2-0c6a785b415e req-0965f6b7-e45e-4552-ae06-e6dcd6714215 service nova] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Received unexpected event network-vif-plugged-c32c52fa-48c3-48d7-87e9-0e9d693d6959 for instance with vm_state building and task_state spawning. [ 827.180380] env[61985]: DEBUG oslo_concurrency.lockutils [req-46670c1b-2981-4d1f-bd07-2adb220582dc req-fd665f4e-ccb1-45db-b657-da8f612a2605 service nova] Releasing lock "refresh_cache-e6a57868-ab6e-45d2-9fa0-94efbb79185b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.192765] env[61985]: DEBUG nova.network.neutron [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Successfully updated port: c32c52fa-48c3-48d7-87e9-0e9d693d6959 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 827.337360] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935829, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.337360] env[61985]: DEBUG nova.compute.manager [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 827.339606] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f94c732d-e53c-45d8-b4ef-83916a88773f tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.341332] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.571s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.341707] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.347057] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.360s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.353242] env[61985]: INFO nova.compute.claims [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.367502] env[61985]: DEBUG nova.network.neutron [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Successfully created port: 876723dc-2a70-4f56-bff6-e8fd73c0a346 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.374972] env[61985]: INFO nova.scheduler.client.report [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Deleted allocations for instance 3815af53-ba3b-4c60-a5fd-2d94102cdd42 [ 827.465975] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935835, 'name': ReconfigVM_Task, 'duration_secs': 0.832555} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.466083] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 761d5954-a5ca-4459-a1d6-bfc59b284bf4/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.466971] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef59de12-821c-4e2d-8cde-78c49254836d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.494116] env[61985]: DEBUG nova.network.neutron [-] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.500729] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a254f81d-e6bb-4116-88c5-228e61947b0b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.517950] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 827.517950] env[61985]: value = "task-935836" [ 827.517950] env[61985]: _type = "Task" [ 827.517950] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.526696] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935836, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.693577] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquiring lock "refresh_cache-402ef006-2835-4239-a3c4-1c37a7bba9a5" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.696355] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquired lock "refresh_cache-402ef006-2835-4239-a3c4-1c37a7bba9a5" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.696355] env[61985]: DEBUG nova.network.neutron [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 827.830213] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935829, 'name': CloneVM_Task, 'duration_secs': 2.109762} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.831396] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Created linked-clone VM from snapshot [ 827.832260] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187cd0a2-8066-4af4-bb28-aabc43cfa6c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.847207] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Uploading image fc3d508b-a533-4ace-90cf-e4f991a60e00 {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 827.879359] env[61985]: DEBUG oslo_vmware.rw_handles [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 827.879359] env[61985]: value = "vm-211342" [ 827.879359] env[61985]: _type = "VirtualMachine" [ 827.879359] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 827.879359] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ff1fdbbf-bf13-4bec-a7f3-45910afc91f9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.887913] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae210b6b-ec63-4797-8b6e-f8feda15ccba tempest-ServerAddressesTestJSON-98582183 tempest-ServerAddressesTestJSON-98582183-project-member] Lock "3815af53-ba3b-4c60-a5fd-2d94102cdd42" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.257s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.892653] env[61985]: DEBUG oslo_vmware.rw_handles [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lease: (returnval){ [ 827.892653] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52084090-1b57-126b-4dff-52bc582b3c22" [ 827.892653] env[61985]: _type = "HttpNfcLease" [ 827.892653] env[61985]: } obtained for exporting VM: (result){ [ 827.892653] env[61985]: value = "vm-211342" [ 827.892653] env[61985]: _type = "VirtualMachine" [ 827.892653] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 827.892997] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the lease: (returnval){ [ 827.892997] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52084090-1b57-126b-4dff-52bc582b3c22" [ 827.892997] env[61985]: _type = "HttpNfcLease" [ 827.892997] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 827.894015] env[61985]: DEBUG nova.compute.manager [req-f6679884-f540-480b-9ca2-0ff549fd97c3 req-56320201-df95-48bb-bce8-487e1307cfbc service nova] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Received event network-vif-deleted-5aa010f5-afab-4ea2-80c4-0d0626458099 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 827.904464] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 827.904464] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52084090-1b57-126b-4dff-52bc582b3c22" [ 827.904464] env[61985]: _type = "HttpNfcLease" [ 827.904464] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 827.905226] env[61985]: DEBUG oslo_vmware.rw_handles [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 827.905226] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52084090-1b57-126b-4dff-52bc582b3c22" [ 827.905226] env[61985]: _type = "HttpNfcLease" [ 827.905226] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 827.905956] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84e44d9-8239-4a16-a339-fc52dba499a2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.914562] env[61985]: DEBUG oslo_vmware.rw_handles [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a7bde-370a-a265-12f2-51ad17880993/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 827.914739] env[61985]: DEBUG oslo_vmware.rw_handles [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a7bde-370a-a265-12f2-51ad17880993/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 828.002031] env[61985]: INFO nova.compute.manager [-] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Took 1.44 seconds to deallocate network for instance. [ 828.008725] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ebc3b27d-7630-4587-81c6-4b07028697e5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.029082] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935836, 'name': ReconfigVM_Task, 'duration_secs': 0.169268} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.031702] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 828.032346] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b6c2007-ded3-4280-8a2a-2e88c38ee70a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.043205] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 828.043205] env[61985]: value = "task-935838" [ 828.043205] env[61985]: _type = "Task" [ 828.043205] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.052439] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935838, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.233700] env[61985]: DEBUG nova.network.neutron [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 828.352145] env[61985]: DEBUG nova.compute.manager [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 828.394303] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.394688] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.394924] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.395240] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.395506] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.395686] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.395970] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.396130] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.396324] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.396503] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.396698] env[61985]: DEBUG nova.virt.hardware [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.398298] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21cf9ca0-427c-47a3-adb2-71ea7b885bd5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.403403] env[61985]: DEBUG nova.network.neutron [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Updating instance_info_cache with network_info: [{"id": "c32c52fa-48c3-48d7-87e9-0e9d693d6959", "address": "fa:16:3e:14:a9:8a", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.79", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32c52fa-48", "ovs_interfaceid": "c32c52fa-48c3-48d7-87e9-0e9d693d6959", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.412261] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78b4172-d4e6-4957-9173-be3f02e48df8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.510538] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.555639] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935838, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.817312] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f481327c-c8dc-4bb3-a381-38441f589579 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.827843] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be918901-0dcc-4c4e-80d8-99bc164dddbe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.869588] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e4b04f-237b-419b-a66c-91586d2f887b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.879032] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cac004d-06ba-4f3c-bdad-ef40953455ec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.895780] env[61985]: DEBUG nova.compute.provider_tree [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.911137] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Releasing lock "refresh_cache-402ef006-2835-4239-a3c4-1c37a7bba9a5" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.911484] env[61985]: DEBUG nova.compute.manager [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Instance network_info: |[{"id": "c32c52fa-48c3-48d7-87e9-0e9d693d6959", "address": "fa:16:3e:14:a9:8a", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.79", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32c52fa-48", "ovs_interfaceid": "c32c52fa-48c3-48d7-87e9-0e9d693d6959", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 828.912202] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:a9:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e1049e8-c06b-4c93-a9e1-2cbb530f3f95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c32c52fa-48c3-48d7-87e9-0e9d693d6959', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.920801] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Creating folder: Project (a25034654f4e411d9d2c895e1412de74). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.921542] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-23c981d8-14d2-474f-b852-ca854e7d81c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.933301] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Created folder: Project (a25034654f4e411d9d2c895e1412de74) in parent group-v211285. [ 828.934071] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Creating folder: Instances. Parent ref: group-v211343. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.934162] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-284a4e1a-0b0c-4542-b5ab-cf8affff19a8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.946438] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Created folder: Instances in parent group-v211343. [ 828.946785] env[61985]: DEBUG oslo.service.loopingcall [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.947453] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 828.947614] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dcdc4d19-3984-4d97-8bfb-2ae15c661a79 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.967645] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.967645] env[61985]: value = "task-935841" [ 828.967645] env[61985]: _type = "Task" [ 828.967645] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.975405] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935841, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.053939] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935838, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.301339] env[61985]: DEBUG oslo_concurrency.lockutils [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "fea0cf39-e851-409f-86f5-31cc128a44dc" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.301688] env[61985]: DEBUG oslo_concurrency.lockutils [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.301982] env[61985]: DEBUG nova.compute.manager [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Going to confirm migration 1 {{(pid=61985) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4789}} [ 829.399548] env[61985]: DEBUG nova.scheduler.client.report [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.478637] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935841, 'name': CreateVM_Task, 'duration_secs': 0.385641} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.478889] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 829.482111] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.482111] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.482111] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.482111] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbb2304f-0fab-4f83-8121-09ac432bf713 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.486057] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for the task: (returnval){ [ 829.486057] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5265aad6-37a3-eb70-100a-90a6ab51c3e5" [ 829.486057] env[61985]: _type = "Task" [ 829.486057] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.493979] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5265aad6-37a3-eb70-100a-90a6ab51c3e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.555558] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935838, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.692650] env[61985]: DEBUG nova.compute.manager [req-9890e786-4aa5-41d2-a8d2-8c9ba0b56fcd req-95f2f2d6-09f7-499b-8613-d2d7455b21aa service nova] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Received event network-changed-c32c52fa-48c3-48d7-87e9-0e9d693d6959 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 829.693066] env[61985]: DEBUG nova.compute.manager [req-9890e786-4aa5-41d2-a8d2-8c9ba0b56fcd req-95f2f2d6-09f7-499b-8613-d2d7455b21aa service nova] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Refreshing instance network info cache due to event network-changed-c32c52fa-48c3-48d7-87e9-0e9d693d6959. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 829.693489] env[61985]: DEBUG oslo_concurrency.lockutils [req-9890e786-4aa5-41d2-a8d2-8c9ba0b56fcd req-95f2f2d6-09f7-499b-8613-d2d7455b21aa service nova] Acquiring lock "refresh_cache-402ef006-2835-4239-a3c4-1c37a7bba9a5" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.693939] env[61985]: DEBUG oslo_concurrency.lockutils [req-9890e786-4aa5-41d2-a8d2-8c9ba0b56fcd req-95f2f2d6-09f7-499b-8613-d2d7455b21aa service nova] Acquired lock "refresh_cache-402ef006-2835-4239-a3c4-1c37a7bba9a5" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.693939] env[61985]: DEBUG nova.network.neutron [req-9890e786-4aa5-41d2-a8d2-8c9ba0b56fcd req-95f2f2d6-09f7-499b-8613-d2d7455b21aa service nova] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Refreshing network info cache for port c32c52fa-48c3-48d7-87e9-0e9d693d6959 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 829.821941] env[61985]: DEBUG nova.network.neutron [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Successfully updated port: 876723dc-2a70-4f56-bff6-e8fd73c0a346 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.904507] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.905085] env[61985]: DEBUG nova.compute.manager [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 829.908875] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.104s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.913020] env[61985]: INFO nova.compute.claims [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.932617] env[61985]: DEBUG oslo_concurrency.lockutils [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.932811] env[61985]: DEBUG oslo_concurrency.lockutils [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.932990] env[61985]: DEBUG nova.network.neutron [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.933189] env[61985]: DEBUG nova.objects.instance [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lazy-loading 'info_cache' on Instance uuid fea0cf39-e851-409f-86f5-31cc128a44dc {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.001313] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5265aad6-37a3-eb70-100a-90a6ab51c3e5, 'name': SearchDatastore_Task, 'duration_secs': 0.013502} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.001805] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.002290] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.002622] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.002898] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.003233] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.003848] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a4fcbe4-81f0-49f6-b21b-b443a7d612c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.013659] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.013972] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 830.015149] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4030b5fa-41f3-4e8d-b8f9-f5ea84c04db9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.023065] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for the task: (returnval){ [ 830.023065] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52179dd3-52a0-05ae-8cdf-645b67dad23e" [ 830.023065] env[61985]: _type = "Task" [ 830.023065] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.036469] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52179dd3-52a0-05ae-8cdf-645b67dad23e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.057712] env[61985]: DEBUG oslo_vmware.api [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935838, 'name': PowerOnVM_Task, 'duration_secs': 1.684576} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.058077] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 830.063806] env[61985]: DEBUG nova.compute.manager [None req-b54691b1-87d8-4c0c-94eb-524ec1a16127 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.067152] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33a2f5f-c032-49ea-8995-9a9479e016b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.124395] env[61985]: DEBUG nova.compute.manager [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Received event network-vif-plugged-876723dc-2a70-4f56-bff6-e8fd73c0a346 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 830.124395] env[61985]: DEBUG oslo_concurrency.lockutils [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] Acquiring lock "c4e95f51-9acc-46e8-9921-e0a85fb38598-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.124395] env[61985]: DEBUG oslo_concurrency.lockutils [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] Lock "c4e95f51-9acc-46e8-9921-e0a85fb38598-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.124395] env[61985]: DEBUG oslo_concurrency.lockutils [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] Lock "c4e95f51-9acc-46e8-9921-e0a85fb38598-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.124395] env[61985]: DEBUG nova.compute.manager [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] No waiting events found dispatching network-vif-plugged-876723dc-2a70-4f56-bff6-e8fd73c0a346 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 830.124606] env[61985]: WARNING nova.compute.manager [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Received unexpected event network-vif-plugged-876723dc-2a70-4f56-bff6-e8fd73c0a346 for instance with vm_state building and task_state spawning. [ 830.124799] env[61985]: DEBUG nova.compute.manager [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Received event network-changed-876723dc-2a70-4f56-bff6-e8fd73c0a346 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 830.125069] env[61985]: DEBUG nova.compute.manager [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Refreshing instance network info cache due to event network-changed-876723dc-2a70-4f56-bff6-e8fd73c0a346. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 830.125352] env[61985]: DEBUG oslo_concurrency.lockutils [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] Acquiring lock "refresh_cache-c4e95f51-9acc-46e8-9921-e0a85fb38598" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.125692] env[61985]: DEBUG oslo_concurrency.lockutils [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] Acquired lock "refresh_cache-c4e95f51-9acc-46e8-9921-e0a85fb38598" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.126313] env[61985]: DEBUG nova.network.neutron [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Refreshing network info cache for port 876723dc-2a70-4f56-bff6-e8fd73c0a346 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 830.326745] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "refresh_cache-c4e95f51-9acc-46e8-9921-e0a85fb38598" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.415208] env[61985]: DEBUG nova.compute.utils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.420431] env[61985]: DEBUG nova.compute.manager [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.420431] env[61985]: DEBUG nova.network.neutron [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 830.430937] env[61985]: DEBUG nova.network.neutron [req-9890e786-4aa5-41d2-a8d2-8c9ba0b56fcd req-95f2f2d6-09f7-499b-8613-d2d7455b21aa service nova] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Updated VIF entry in instance network info cache for port c32c52fa-48c3-48d7-87e9-0e9d693d6959. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 830.431377] env[61985]: DEBUG nova.network.neutron [req-9890e786-4aa5-41d2-a8d2-8c9ba0b56fcd req-95f2f2d6-09f7-499b-8613-d2d7455b21aa service nova] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Updating instance_info_cache with network_info: [{"id": "c32c52fa-48c3-48d7-87e9-0e9d693d6959", "address": "fa:16:3e:14:a9:8a", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.79", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc32c52fa-48", "ovs_interfaceid": "c32c52fa-48c3-48d7-87e9-0e9d693d6959", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.494721] env[61985]: DEBUG nova.policy [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a78dbc667d24c819e6447740fc2d8e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2dbceebf07c43ac90ed80f8a09294b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.538178] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52179dd3-52a0-05ae-8cdf-645b67dad23e, 'name': SearchDatastore_Task, 'duration_secs': 0.0135} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.539053] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db6b2f60-96cb-46e2-b0fa-62fc504b14a9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.545584] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for the task: (returnval){ [ 830.545584] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d96882-b87c-74f0-98ee-fc98ae4580ca" [ 830.545584] env[61985]: _type = "Task" [ 830.545584] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.555298] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d96882-b87c-74f0-98ee-fc98ae4580ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.672065] env[61985]: DEBUG nova.network.neutron [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.762386] env[61985]: DEBUG nova.network.neutron [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.920918] env[61985]: DEBUG nova.compute.manager [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 830.934092] env[61985]: DEBUG oslo_concurrency.lockutils [req-9890e786-4aa5-41d2-a8d2-8c9ba0b56fcd req-95f2f2d6-09f7-499b-8613-d2d7455b21aa service nova] Releasing lock "refresh_cache-402ef006-2835-4239-a3c4-1c37a7bba9a5" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.060697] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d96882-b87c-74f0-98ee-fc98ae4580ca, 'name': SearchDatastore_Task, 'duration_secs': 0.013184} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.063285] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.063571] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 402ef006-2835-4239-a3c4-1c37a7bba9a5/402ef006-2835-4239-a3c4-1c37a7bba9a5.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 831.063911] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a02ed5f-d271-41d0-a403-74daca0c0e6d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.073517] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for the task: (returnval){ [ 831.073517] env[61985]: value = "task-935842" [ 831.073517] env[61985]: _type = "Task" [ 831.073517] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.083697] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935842, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.239253] env[61985]: DEBUG nova.network.neutron [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Successfully created port: 4959d778-7709-41bf-a558-f6ba7cce7d4b {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.266280] env[61985]: DEBUG oslo_concurrency.lockutils [req-3711e6f5-0d6d-47dd-ad42-e173a4e6ab2d req-93427987-64cc-41bc-99a8-7c574678589c service nova] Releasing lock "refresh_cache-c4e95f51-9acc-46e8-9921-e0a85fb38598" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.266280] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "refresh_cache-c4e95f51-9acc-46e8-9921-e0a85fb38598" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.266280] env[61985]: DEBUG nova.network.neutron [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 831.312410] env[61985]: DEBUG nova.network.neutron [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance_info_cache with network_info: [{"id": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "address": "fa:16:3e:7b:a1:a8", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadde3d48-eb", "ovs_interfaceid": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.345808] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79d06be-fb21-4071-afdb-9752afaeb214 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.357409] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7085a859-f081-44ad-af09-d9f76fc0aec1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.394549] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6102f2-f276-4ceb-83ec-3eaae90d5020 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.403697] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd27419-22e8-43f3-967d-e78be494df83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.418407] env[61985]: DEBUG nova.compute.provider_tree [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.583928] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935842, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.813121] env[61985]: DEBUG nova.network.neutron [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 831.816244] env[61985]: DEBUG oslo_concurrency.lockutils [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.816668] env[61985]: DEBUG nova.objects.instance [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lazy-loading 'migration_context' on Instance uuid fea0cf39-e851-409f-86f5-31cc128a44dc {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.924565] env[61985]: DEBUG nova.scheduler.client.report [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.936945] env[61985]: DEBUG nova.compute.manager [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 831.954191] env[61985]: INFO nova.compute.manager [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Unrescuing [ 831.954466] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.954626] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.954797] env[61985]: DEBUG nova.network.neutron [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 831.967957] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.969535] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.969535] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.969535] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.969535] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.969535] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.971662] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.971662] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.971662] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.971662] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.971662] env[61985]: DEBUG nova.virt.hardware [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.971939] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18540c9b-2bc8-4749-94bb-c35996e4c03b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.982697] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a411c210-7672-49c2-9f64-fbee636ba0c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.990094] env[61985]: DEBUG nova.network.neutron [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Updating instance_info_cache with network_info: [{"id": "876723dc-2a70-4f56-bff6-e8fd73c0a346", "address": "fa:16:3e:4f:ef:32", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap876723dc-2a", "ovs_interfaceid": "876723dc-2a70-4f56-bff6-e8fd73c0a346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.087395] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935842, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651175} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.088443] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 402ef006-2835-4239-a3c4-1c37a7bba9a5/402ef006-2835-4239-a3c4-1c37a7bba9a5.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 832.090976] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.090976] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03936f78-1fd4-4a79-83e1-801d583f5979 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.095748] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for the task: (returnval){ [ 832.095748] env[61985]: value = "task-935843" [ 832.095748] env[61985]: _type = "Task" [ 832.095748] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.104577] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.323041] env[61985]: DEBUG nova.objects.base [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 832.323041] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4ab43b-8a9a-4f45-9b51-c08f1a2e2e5a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.344935] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8eadc7d2-00d1-4aa4-9ef8-ce3f80443b61 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.351321] env[61985]: DEBUG oslo_vmware.api [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 832.351321] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525bca92-cbb8-bb1c-53c2-515f4306db6a" [ 832.351321] env[61985]: _type = "Task" [ 832.351321] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.361787] env[61985]: DEBUG oslo_vmware.api [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525bca92-cbb8-bb1c-53c2-515f4306db6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.428031] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.428673] env[61985]: DEBUG nova.compute.manager [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 832.433041] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.468s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.434862] env[61985]: INFO nova.compute.claims [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.502028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "refresh_cache-c4e95f51-9acc-46e8-9921-e0a85fb38598" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.502257] env[61985]: DEBUG nova.compute.manager [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Instance network_info: |[{"id": "876723dc-2a70-4f56-bff6-e8fd73c0a346", "address": "fa:16:3e:4f:ef:32", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap876723dc-2a", "ovs_interfaceid": "876723dc-2a70-4f56-bff6-e8fd73c0a346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 832.502712] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:ef:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2c019b6-3ef3-4c8f-95bd-edede2c554a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '876723dc-2a70-4f56-bff6-e8fd73c0a346', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.510414] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating folder: Project (2e66b7a6e7354b06a77295eadf4e5e0f). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 832.511072] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4778961a-8d56-46df-b010-7994257a6fd9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.523898] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Created folder: Project (2e66b7a6e7354b06a77295eadf4e5e0f) in parent group-v211285. [ 832.524116] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating folder: Instances. Parent ref: group-v211346. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 832.524358] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d37b5829-c73e-4126-9179-37d1dc872197 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.533733] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Created folder: Instances in parent group-v211346. [ 832.534290] env[61985]: DEBUG oslo.service.loopingcall [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.534290] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 832.534447] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-177486d7-5c13-4004-96fe-e1cac14a19ce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.557171] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.557171] env[61985]: value = "task-935846" [ 832.557171] env[61985]: _type = "Task" [ 832.557171] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.567117] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935846, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.607075] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07971} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.608692] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.609553] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2830b26-91b6-4aa3-818e-f2e729279b11 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.634213] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 402ef006-2835-4239-a3c4-1c37a7bba9a5/402ef006-2835-4239-a3c4-1c37a7bba9a5.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.634569] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28292c01-b108-403e-aaa0-0e23c878b763 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.657591] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for the task: (returnval){ [ 832.657591] env[61985]: value = "task-935847" [ 832.657591] env[61985]: _type = "Task" [ 832.657591] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.665951] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935847, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.800191] env[61985]: DEBUG nova.network.neutron [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Updating instance_info_cache with network_info: [{"id": "6ea4b2df-e4c4-48a4-9af2-fe4606759004", "address": "fa:16:3e:b3:50:b6", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ea4b2df-e4", "ovs_interfaceid": "6ea4b2df-e4c4-48a4-9af2-fe4606759004", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.862413] env[61985]: DEBUG oslo_vmware.api [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525bca92-cbb8-bb1c-53c2-515f4306db6a, 'name': SearchDatastore_Task, 'duration_secs': 0.01063} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.862769] env[61985]: DEBUG oslo_concurrency.lockutils [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.941026] env[61985]: DEBUG nova.compute.utils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.943097] env[61985]: DEBUG nova.compute.manager [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.943417] env[61985]: DEBUG nova.network.neutron [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 832.992120] env[61985]: DEBUG nova.policy [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '149807bc94cc4cddb59e4dad0bf1099c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8015ef21eaf4a0bb19063e261064cdc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.072017] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935846, 'name': CreateVM_Task, 'duration_secs': 0.443261} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.072017] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 833.072017] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.072017] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.072017] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 833.072017] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16129588-8569-40e2-8259-5569f0bbd18b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.075752] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 833.075752] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ff2a50-30c6-049b-4444-58daf52f61b0" [ 833.075752] env[61985]: _type = "Task" [ 833.075752] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.084669] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ff2a50-30c6-049b-4444-58daf52f61b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.170157] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935847, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.248247] env[61985]: DEBUG nova.network.neutron [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Successfully created port: 5b8f2f2c-0511-4436-982f-3f32c10f5759 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.303540] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "refresh_cache-761d5954-a5ca-4459-a1d6-bfc59b284bf4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.304249] env[61985]: DEBUG nova.objects.instance [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lazy-loading 'flavor' on Instance uuid 761d5954-a5ca-4459-a1d6-bfc59b284bf4 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.445176] env[61985]: DEBUG nova.compute.manager [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 833.491265] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.491731] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.491982] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.492147] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Cleaning up deleted instances {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11232}} [ 833.588689] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ff2a50-30c6-049b-4444-58daf52f61b0, 'name': SearchDatastore_Task, 'duration_secs': 0.018515} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.588689] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.588689] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.588689] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.588850] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.588850] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.589299] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7c9ddd4-60c7-4b49-ba82-587cd40e2754 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.604050] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.604050] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 833.604339] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64a110b7-3a9b-4299-97c9-e99c01eea92e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.611022] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 833.611022] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f1ca91-856f-7e21-9e50-27eae3985be5" [ 833.611022] env[61985]: _type = "Task" [ 833.611022] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.626058] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f1ca91-856f-7e21-9e50-27eae3985be5, 'name': SearchDatastore_Task, 'duration_secs': 0.010976} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.627658] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9c551a0-4b46-49c4-9734-e464d5be150c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.633135] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 833.633135] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a0b54a-2086-9e7e-5f55-e5415d50eae6" [ 833.633135] env[61985]: _type = "Task" [ 833.633135] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.644386] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a0b54a-2086-9e7e-5f55-e5415d50eae6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.669506] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935847, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.816053] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb0ba22-97fe-4e33-8745-20dacfe7c2b2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.835454] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 833.839483] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a18378fe-81b8-4073-924b-d3688a4a64fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.849859] env[61985]: DEBUG oslo_vmware.api [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 833.849859] env[61985]: value = "task-935848" [ 833.849859] env[61985]: _type = "Task" [ 833.849859] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.864288] env[61985]: DEBUG oslo_vmware.api [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935848, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.878991] env[61985]: DEBUG nova.network.neutron [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Successfully updated port: 4959d778-7709-41bf-a558-f6ba7cce7d4b {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 833.943885] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8b6113-8fcc-44e1-92c7-8a4922d4d658 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.955531] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5db7df8-3231-48e2-9b2b-ad77218d4851 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.999423] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c7116a-4b5d-4202-b10c-dd31f23b4b25 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.999933] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] There are 5 instances to clean {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 833.999933] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 3815af53-ba3b-4c60-a5fd-2d94102cdd42] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 834.006695] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c6ea4f-0400-4c71-a92d-504e461dd85b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.022190] env[61985]: DEBUG nova.compute.provider_tree [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.147980] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a0b54a-2086-9e7e-5f55-e5415d50eae6, 'name': SearchDatastore_Task, 'duration_secs': 0.014281} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.148346] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.148716] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] c4e95f51-9acc-46e8-9921-e0a85fb38598/c4e95f51-9acc-46e8-9921-e0a85fb38598.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 834.148979] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79574368-9531-498a-9481-b863f0c98cc0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.157459] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 834.157459] env[61985]: value = "task-935849" [ 834.157459] env[61985]: _type = "Task" [ 834.157459] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.171034] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935847, 'name': ReconfigVM_Task, 'duration_secs': 1.426884} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.173115] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 402ef006-2835-4239-a3c4-1c37a7bba9a5/402ef006-2835-4239-a3c4-1c37a7bba9a5.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.173756] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935849, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.174023] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0de7d882-90d6-4a21-a3cc-52fbd5b795c2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.182142] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for the task: (returnval){ [ 834.182142] env[61985]: value = "task-935850" [ 834.182142] env[61985]: _type = "Task" [ 834.182142] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.194817] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935850, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.280305] env[61985]: DEBUG nova.compute.manager [req-e7426b77-5ecf-48b3-8af6-2bfed06e2c4e req-58019fbe-6593-426e-949f-cffe5fb7303b service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received event network-vif-plugged-4959d778-7709-41bf-a558-f6ba7cce7d4b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 834.280572] env[61985]: DEBUG oslo_concurrency.lockutils [req-e7426b77-5ecf-48b3-8af6-2bfed06e2c4e req-58019fbe-6593-426e-949f-cffe5fb7303b service nova] Acquiring lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.280819] env[61985]: DEBUG oslo_concurrency.lockutils [req-e7426b77-5ecf-48b3-8af6-2bfed06e2c4e req-58019fbe-6593-426e-949f-cffe5fb7303b service nova] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.281210] env[61985]: DEBUG oslo_concurrency.lockutils [req-e7426b77-5ecf-48b3-8af6-2bfed06e2c4e req-58019fbe-6593-426e-949f-cffe5fb7303b service nova] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.281439] env[61985]: DEBUG nova.compute.manager [req-e7426b77-5ecf-48b3-8af6-2bfed06e2c4e req-58019fbe-6593-426e-949f-cffe5fb7303b service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] No waiting events found dispatching network-vif-plugged-4959d778-7709-41bf-a558-f6ba7cce7d4b {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 834.281648] env[61985]: WARNING nova.compute.manager [req-e7426b77-5ecf-48b3-8af6-2bfed06e2c4e req-58019fbe-6593-426e-949f-cffe5fb7303b service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received unexpected event network-vif-plugged-4959d778-7709-41bf-a558-f6ba7cce7d4b for instance with vm_state building and task_state spawning. [ 834.360333] env[61985]: DEBUG oslo_vmware.api [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935848, 'name': PowerOffVM_Task, 'duration_secs': 0.348351} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.360333] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 834.367833] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Reconfiguring VM instance instance-0000002d to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 834.368923] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b86ead39-d079-457a-8590-3f482d431cba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.391571] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.391571] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.391571] env[61985]: DEBUG nova.network.neutron [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 834.405519] env[61985]: DEBUG oslo_vmware.api [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 834.405519] env[61985]: value = "task-935851" [ 834.405519] env[61985]: _type = "Task" [ 834.405519] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.416058] env[61985]: DEBUG oslo_vmware.api [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935851, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.462217] env[61985]: DEBUG nova.compute.manager [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 834.511944] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.512493] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.512758] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.513117] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.513397] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.513668] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.514429] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.514744] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.515057] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.515401] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.515714] env[61985]: DEBUG nova.virt.hardware [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.516363] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 4add2608-8187-42a2-9897-3e5a9049757e] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 834.519888] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26faed31-bf0e-4916-8c97-22b022f22c3b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.526231] env[61985]: DEBUG nova.network.neutron [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 834.531607] env[61985]: DEBUG nova.scheduler.client.report [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.542019] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537b4545-10cd-4de1-906a-eab1206bb458 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.668493] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935849, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50566} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.672021] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] c4e95f51-9acc-46e8-9921-e0a85fb38598/c4e95f51-9acc-46e8-9921-e0a85fb38598.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 834.672021] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.672021] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a58832de-33db-4530-b807-ab2e24dd86f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.676335] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 834.676335] env[61985]: value = "task-935852" [ 834.676335] env[61985]: _type = "Task" [ 834.676335] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.690189] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935852, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.699148] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935850, 'name': Rename_Task, 'duration_secs': 0.172227} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.699148] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 834.699148] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e920537-c8fc-4135-868d-0d7280ba82ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.707196] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for the task: (returnval){ [ 834.707196] env[61985]: value = "task-935853" [ 834.707196] env[61985]: _type = "Task" [ 834.707196] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.717488] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935853, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.829665] env[61985]: DEBUG nova.network.neutron [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Successfully updated port: 5b8f2f2c-0511-4436-982f-3f32c10f5759 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.916890] env[61985]: DEBUG oslo_vmware.api [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935851, 'name': ReconfigVM_Task, 'duration_secs': 0.258056} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.917203] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Reconfigured VM instance instance-0000002d to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 834.917472] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 834.917801] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c55e0c39-8984-49ab-90a0-0cbbc71993bf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.926029] env[61985]: DEBUG oslo_vmware.api [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 834.926029] env[61985]: value = "task-935854" [ 834.926029] env[61985]: _type = "Task" [ 834.926029] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.935742] env[61985]: DEBUG oslo_vmware.api [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935854, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.026315] env[61985]: DEBUG nova.network.neutron [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updating instance_info_cache with network_info: [{"id": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "address": "fa:16:3e:68:94:dd", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4959d778-77", "ovs_interfaceid": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.031169] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 2e1e46ea-477c-41f3-938f-bd654bc5999d] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 835.036812] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.604s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.037433] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 835.040643] env[61985]: DEBUG oslo_concurrency.lockutils [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.436s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.040840] env[61985]: DEBUG nova.objects.instance [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61985) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 835.051043] env[61985]: DEBUG nova.compute.manager [req-44386262-5da6-4b63-a221-54032fbf452c req-b747a7ab-77f3-47a9-aafb-7b4e0e8270a1 service nova] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Received event network-vif-plugged-5b8f2f2c-0511-4436-982f-3f32c10f5759 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 835.051351] env[61985]: DEBUG oslo_concurrency.lockutils [req-44386262-5da6-4b63-a221-54032fbf452c req-b747a7ab-77f3-47a9-aafb-7b4e0e8270a1 service nova] Acquiring lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.051680] env[61985]: DEBUG oslo_concurrency.lockutils [req-44386262-5da6-4b63-a221-54032fbf452c req-b747a7ab-77f3-47a9-aafb-7b4e0e8270a1 service nova] Lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.051890] env[61985]: DEBUG oslo_concurrency.lockutils [req-44386262-5da6-4b63-a221-54032fbf452c req-b747a7ab-77f3-47a9-aafb-7b4e0e8270a1 service nova] Lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.052306] env[61985]: DEBUG nova.compute.manager [req-44386262-5da6-4b63-a221-54032fbf452c req-b747a7ab-77f3-47a9-aafb-7b4e0e8270a1 service nova] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] No waiting events found dispatching network-vif-plugged-5b8f2f2c-0511-4436-982f-3f32c10f5759 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.052363] env[61985]: WARNING nova.compute.manager [req-44386262-5da6-4b63-a221-54032fbf452c req-b747a7ab-77f3-47a9-aafb-7b4e0e8270a1 service nova] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Received unexpected event network-vif-plugged-5b8f2f2c-0511-4436-982f-3f32c10f5759 for instance with vm_state building and task_state spawning. [ 835.187545] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935852, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089522} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.187835] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.188673] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3eb7f95-f801-424d-b3f1-646a6a626890 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.211495] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] c4e95f51-9acc-46e8-9921-e0a85fb38598/c4e95f51-9acc-46e8-9921-e0a85fb38598.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.211903] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d208a4dc-d5fc-4500-90d7-0a8f8fac2e01 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.236615] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935853, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.238358] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 835.238358] env[61985]: value = "task-935855" [ 835.238358] env[61985]: _type = "Task" [ 835.238358] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.247401] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935855, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.333160] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "refresh_cache-83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.333305] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired lock "refresh_cache-83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.333462] env[61985]: DEBUG nova.network.neutron [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 835.438286] env[61985]: DEBUG oslo_vmware.api [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935854, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.528921] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.529580] env[61985]: DEBUG nova.compute.manager [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Instance network_info: |[{"id": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "address": "fa:16:3e:68:94:dd", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4959d778-77", "ovs_interfaceid": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 835.530221] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:94:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4959d778-7709-41bf-a558-f6ba7cce7d4b', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.543293] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Creating folder: Project (a2dbceebf07c43ac90ed80f8a09294b7). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 835.544084] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 213a3e19-5589-4261-96b0-69acfb6798ef] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 835.547764] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-07391723-cf80-475e-9538-7dc9faaa51df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.553758] env[61985]: DEBUG nova.compute.utils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.561073] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 835.561612] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 835.574491] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Created folder: Project (a2dbceebf07c43ac90ed80f8a09294b7) in parent group-v211285. [ 835.575053] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Creating folder: Instances. Parent ref: group-v211349. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 835.576463] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5c74a15-498a-4dec-b6d1-2f9154c4bfd4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.589575] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Created folder: Instances in parent group-v211349. [ 835.589889] env[61985]: DEBUG oslo.service.loopingcall [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.590288] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 835.590746] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dfb20ceb-074f-4b4d-8955-3341b981a8c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.611393] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.611393] env[61985]: value = "task-935858" [ 835.611393] env[61985]: _type = "Task" [ 835.611393] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.620961] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935858, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.724039] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935853, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.729151] env[61985]: DEBUG nova.policy [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3d8fab40f554eb3a4804a4291686146', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c06fc8ffa004e00be0a1991d2e082e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 835.748712] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935855, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.876096] env[61985]: DEBUG nova.network.neutron [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 835.939649] env[61985]: DEBUG oslo_vmware.api [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-935854, 'name': PowerOnVM_Task, 'duration_secs': 0.62494} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.940768] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 835.940768] env[61985]: DEBUG nova.compute.manager [None req-e61a4a23-89dd-43e1-a452-10fa7606353c tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.941248] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03cf2248-7fa6-495a-9b7e-22481980cd36 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.052191] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 8eb535d6-b993-4453-ade0-ed6c36125690] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 836.064036] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 836.066845] env[61985]: DEBUG oslo_concurrency.lockutils [None req-570de065-c6eb-4084-8e60-bb79429590ea tempest-ServersAdmin275Test-940564502 tempest-ServersAdmin275Test-940564502-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.026s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.068741] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.143s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.070890] env[61985]: INFO nova.compute.claims [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.076185] env[61985]: DEBUG nova.network.neutron [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Updating instance_info_cache with network_info: [{"id": "5b8f2f2c-0511-4436-982f-3f32c10f5759", "address": "fa:16:3e:ff:c2:cf", "network": {"id": "ddf63d11-2633-4be8-b73f-c5ca846b420c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-709607861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8015ef21eaf4a0bb19063e261064cdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b8f2f2c-05", "ovs_interfaceid": "5b8f2f2c-0511-4436-982f-3f32c10f5759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.125485] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935858, 'name': CreateVM_Task, 'duration_secs': 0.46025} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.125485] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 836.126204] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.126437] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.126845] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.128914] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70a4e62b-a8f5-43c1-9a88-90d08df364c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.135862] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 836.135862] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a54f1e-b350-6d9f-99c8-d53e5ffcc3ab" [ 836.135862] env[61985]: _type = "Task" [ 836.135862] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.145700] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a54f1e-b350-6d9f-99c8-d53e5ffcc3ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.226532] env[61985]: DEBUG oslo_vmware.api [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935853, 'name': PowerOnVM_Task, 'duration_secs': 1.410194} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.227016] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 836.227298] env[61985]: INFO nova.compute.manager [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Took 10.52 seconds to spawn the instance on the hypervisor. [ 836.227667] env[61985]: DEBUG nova.compute.manager [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.228526] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b35034f-5efd-4cb0-8e28-947bbbd07315 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.249256] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935855, 'name': ReconfigVM_Task, 'duration_secs': 0.57474} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.249256] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Reconfigured VM instance instance-00000034 to attach disk [datastore1] c4e95f51-9acc-46e8-9921-e0a85fb38598/c4e95f51-9acc-46e8-9921-e0a85fb38598.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.249256] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3f5ecc95-4f8d-46a8-9f19-0be127629a20 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.255753] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 836.255753] env[61985]: value = "task-935859" [ 836.255753] env[61985]: _type = "Task" [ 836.255753] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.270592] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935859, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.555809] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 836.556090] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Cleaning up deleted instances with incomplete migration {{(pid=61985) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11270}} [ 836.581163] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Releasing lock "refresh_cache-83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.581163] env[61985]: DEBUG nova.compute.manager [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Instance network_info: |[{"id": "5b8f2f2c-0511-4436-982f-3f32c10f5759", "address": "fa:16:3e:ff:c2:cf", "network": {"id": "ddf63d11-2633-4be8-b73f-c5ca846b420c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-709607861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8015ef21eaf4a0bb19063e261064cdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b8f2f2c-05", "ovs_interfaceid": "5b8f2f2c-0511-4436-982f-3f32c10f5759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 836.588022] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:c2:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1470a3f8-be8a-4339-8a6f-9519366f32e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5b8f2f2c-0511-4436-982f-3f32c10f5759', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.601943] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Creating folder: Project (a8015ef21eaf4a0bb19063e261064cdc). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 836.604060] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e1b9f519-e5a7-48fe-81c1-620861df336e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.618016] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Created folder: Project (a8015ef21eaf4a0bb19063e261064cdc) in parent group-v211285. [ 836.618416] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Creating folder: Instances. Parent ref: group-v211352. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 836.618945] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c7bbdcb-1627-483f-ac40-4806ff49a63b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.628135] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Created folder: Instances in parent group-v211352. [ 836.628330] env[61985]: DEBUG oslo.service.loopingcall [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.629260] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 836.629534] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-18b3e27c-bff7-45b7-9b69-020531726a87 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.658387] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a54f1e-b350-6d9f-99c8-d53e5ffcc3ab, 'name': SearchDatastore_Task, 'duration_secs': 0.042189} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.659996] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.660352] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.660616] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.660763] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.661014] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.661277] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.661277] env[61985]: value = "task-935862" [ 836.661277] env[61985]: _type = "Task" [ 836.661277] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.661484] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1bc6fe41-2f19-4e77-b6a0-44c4ccb3264c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.677029] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935862, 'name': CreateVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.678587] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.678834] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 836.679614] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a3f21cc-959a-4a61-936c-44b971682c1a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.689099] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 836.689099] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d62af2-8048-1258-0558-dde004f98743" [ 836.689099] env[61985]: _type = "Task" [ 836.689099] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.696242] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d62af2-8048-1258-0558-dde004f98743, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.754835] env[61985]: INFO nova.compute.manager [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Took 39.77 seconds to build instance. [ 836.765900] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935859, 'name': Rename_Task, 'duration_secs': 0.221556} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.768699] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.768699] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1f94f81-92ad-4397-96a0-1931196a8a6f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.774122] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 836.774122] env[61985]: value = "task-935863" [ 836.774122] env[61985]: _type = "Task" [ 836.774122] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.787862] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.900484] env[61985]: DEBUG nova.compute.manager [req-25c21b1a-ad38-45c3-bc0f-71067826c58a req-eb6506dd-9ff1-4012-ba0a-661edfe13d32 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received event network-changed-4959d778-7709-41bf-a558-f6ba7cce7d4b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 836.900703] env[61985]: DEBUG nova.compute.manager [req-25c21b1a-ad38-45c3-bc0f-71067826c58a req-eb6506dd-9ff1-4012-ba0a-661edfe13d32 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Refreshing instance network info cache due to event network-changed-4959d778-7709-41bf-a558-f6ba7cce7d4b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 836.900936] env[61985]: DEBUG oslo_concurrency.lockutils [req-25c21b1a-ad38-45c3-bc0f-71067826c58a req-eb6506dd-9ff1-4012-ba0a-661edfe13d32 service nova] Acquiring lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.901641] env[61985]: DEBUG oslo_concurrency.lockutils [req-25c21b1a-ad38-45c3-bc0f-71067826c58a req-eb6506dd-9ff1-4012-ba0a-661edfe13d32 service nova] Acquired lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.901879] env[61985]: DEBUG nova.network.neutron [req-25c21b1a-ad38-45c3-bc0f-71067826c58a req-eb6506dd-9ff1-4012-ba0a-661edfe13d32 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Refreshing network info cache for port 4959d778-7709-41bf-a558-f6ba7cce7d4b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 836.915568] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Successfully created port: d3b72151-261d-424c-adf2-aa8c1b98ec5e {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.060561] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.079037] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 837.111118] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.111118] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.111118] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.111399] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.111399] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.111678] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.112056] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.112353] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.112674] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.112985] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.113315] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.114361] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99265363-70c4-4dc5-932a-f200373a40d5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.129023] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6fc3af-8ce9-45ce-8066-911206ed82ca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.176200] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935862, 'name': CreateVM_Task, 'duration_secs': 0.39619} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.176607] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 837.177465] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.177804] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.178257] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.178670] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eff554b6-0a2f-4283-a1e3-cec821f7541f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.184833] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 837.184833] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52363f3b-6d83-eee6-abc2-c11984a897e1" [ 837.184833] env[61985]: _type = "Task" [ 837.184833] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.200954] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d62af2-8048-1258-0558-dde004f98743, 'name': SearchDatastore_Task, 'duration_secs': 0.012701} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.206485] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52363f3b-6d83-eee6-abc2-c11984a897e1, 'name': SearchDatastore_Task, 'duration_secs': 0.010852} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.209313] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd2d744f-2d8d-449b-87c6-f449aa7bd9ce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.212704] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.212704] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.212704] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.219025] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 837.219025] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52217bdd-346c-09a4-e6f5-b95d84306ad5" [ 837.219025] env[61985]: _type = "Task" [ 837.219025] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.227280] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52217bdd-346c-09a4-e6f5-b95d84306ad5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.260022] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20cecdd4-101c-45a4-ba1e-53d3a63822b9 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lock "402ef006-2835-4239-a3c4-1c37a7bba9a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.815s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.296353] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935863, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.540034] env[61985]: DEBUG oslo_vmware.rw_handles [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a7bde-370a-a265-12f2-51ad17880993/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 837.540930] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb97d8f-a2b7-4262-a0f0-2301849728d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.551609] env[61985]: DEBUG oslo_vmware.rw_handles [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a7bde-370a-a265-12f2-51ad17880993/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 837.552122] env[61985]: ERROR oslo_vmware.rw_handles [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a7bde-370a-a265-12f2-51ad17880993/disk-0.vmdk due to incomplete transfer. [ 837.552122] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4d7f8630-f457-4921-b2db-f47a79dcac7e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.563526] env[61985]: DEBUG oslo_vmware.rw_handles [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524a7bde-370a-a265-12f2-51ad17880993/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 837.563526] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Uploaded image fc3d508b-a533-4ace-90cf-e4f991a60e00 to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 837.564050] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 837.564469] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-cab94c47-78c1-4721-bb90-85228b5d772a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.572074] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 837.572074] env[61985]: value = "task-935864" [ 837.572074] env[61985]: _type = "Task" [ 837.572074] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.595037] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935864, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.664679] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-631d97f2-d096-4a05-8f7e-93330063ea48 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.675891] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b489f45-f4d0-483c-8fee-28b120cfeff8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.680695] env[61985]: DEBUG nova.network.neutron [req-25c21b1a-ad38-45c3-bc0f-71067826c58a req-eb6506dd-9ff1-4012-ba0a-661edfe13d32 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updated VIF entry in instance network info cache for port 4959d778-7709-41bf-a558-f6ba7cce7d4b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 837.681167] env[61985]: DEBUG nova.network.neutron [req-25c21b1a-ad38-45c3-bc0f-71067826c58a req-eb6506dd-9ff1-4012-ba0a-661edfe13d32 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updating instance_info_cache with network_info: [{"id": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "address": "fa:16:3e:68:94:dd", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4959d778-77", "ovs_interfaceid": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.720166] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310a780b-851a-426d-b6b4-727e46814c6c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.738398] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52217bdd-346c-09a4-e6f5-b95d84306ad5, 'name': SearchDatastore_Task, 'duration_secs': 0.010546} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.741010] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.741336] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf/5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 837.741705] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.741990] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.742239] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e979c1d-4a31-49b0-a6d5-09dccfea18c2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.745641] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79074b94-096b-4b35-98d6-89e57aa08702 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.749604] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cdc6012c-e6cc-4aaf-9e68-46d4de903d1e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.754857] env[61985]: DEBUG nova.compute.manager [None req-2242c50e-b5e2-44a5-a84f-dd469067e49a tempest-ServerDiagnosticsTest-1518075228 tempest-ServerDiagnosticsTest-1518075228-project-admin] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.756423] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfa40c9-eaca-4ccc-8ee2-c2f4d7640395 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.770446] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 837.770446] env[61985]: value = "task-935865" [ 837.770446] env[61985]: _type = "Task" [ 837.770446] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.770446] env[61985]: DEBUG nova.compute.manager [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 837.773122] env[61985]: DEBUG nova.compute.provider_tree [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.776374] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.776561] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 837.784099] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe7ad61d-b27e-4f23-b0a7-f567cc207341 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.788648] env[61985]: INFO nova.compute.manager [None req-2242c50e-b5e2-44a5-a84f-dd469067e49a tempest-ServerDiagnosticsTest-1518075228 tempest-ServerDiagnosticsTest-1518075228-project-admin] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Retrieving diagnostics [ 837.790488] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7edb0b39-7591-4d27-b75e-d2f64f99ce8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.801152] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 837.801152] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b37e16-101f-3d49-5ca8-a914690ea5dd" [ 837.801152] env[61985]: _type = "Task" [ 837.801152] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.801462] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935865, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.801756] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935863, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.841361] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b37e16-101f-3d49-5ca8-a914690ea5dd, 'name': SearchDatastore_Task, 'duration_secs': 0.011239} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.842226] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34347720-95d5-4867-91f4-913784db810d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.848422] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 837.848422] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520cd259-8f04-865b-f9a5-582f54db7aac" [ 837.848422] env[61985]: _type = "Task" [ 837.848422] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.857740] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520cd259-8f04-865b-f9a5-582f54db7aac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.923094] env[61985]: DEBUG nova.compute.manager [req-ee98f274-153c-4dd7-a029-ac47d46d60e8 req-c864b820-6b10-4a61-8892-187f539782fe service nova] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Received event network-changed-5b8f2f2c-0511-4436-982f-3f32c10f5759 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 837.923242] env[61985]: DEBUG nova.compute.manager [req-ee98f274-153c-4dd7-a029-ac47d46d60e8 req-c864b820-6b10-4a61-8892-187f539782fe service nova] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Refreshing instance network info cache due to event network-changed-5b8f2f2c-0511-4436-982f-3f32c10f5759. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 837.923546] env[61985]: DEBUG oslo_concurrency.lockutils [req-ee98f274-153c-4dd7-a029-ac47d46d60e8 req-c864b820-6b10-4a61-8892-187f539782fe service nova] Acquiring lock "refresh_cache-83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.923659] env[61985]: DEBUG oslo_concurrency.lockutils [req-ee98f274-153c-4dd7-a029-ac47d46d60e8 req-c864b820-6b10-4a61-8892-187f539782fe service nova] Acquired lock "refresh_cache-83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.923796] env[61985]: DEBUG nova.network.neutron [req-ee98f274-153c-4dd7-a029-ac47d46d60e8 req-c864b820-6b10-4a61-8892-187f539782fe service nova] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Refreshing network info cache for port 5b8f2f2c-0511-4436-982f-3f32c10f5759 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 838.083222] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935864, 'name': Destroy_Task, 'duration_secs': 0.338753} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.083544] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Destroyed the VM [ 838.083891] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 838.084186] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0fadca7c-9a9a-445a-924b-7b1a96e01c7a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.093444] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 838.093444] env[61985]: value = "task-935866" [ 838.093444] env[61985]: _type = "Task" [ 838.093444] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.104519] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935866, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.184240] env[61985]: DEBUG oslo_concurrency.lockutils [req-25c21b1a-ad38-45c3-bc0f-71067826c58a req-eb6506dd-9ff1-4012-ba0a-661edfe13d32 service nova] Releasing lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.279095] env[61985]: DEBUG nova.scheduler.client.report [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.307256] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935865, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490432} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.309069] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf/5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 838.309499] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.310118] env[61985]: DEBUG oslo_vmware.api [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935863, 'name': PowerOnVM_Task, 'duration_secs': 1.265476} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.310543] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8bd98c3d-802b-403a-9235-6b5a38b495a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.314854] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 838.314854] env[61985]: INFO nova.compute.manager [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Took 9.96 seconds to spawn the instance on the hypervisor. [ 838.314854] env[61985]: DEBUG nova.compute.manager [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.315672] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b966e8e-13f2-4cf6-86f5-f12db970f370 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.328733] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.334676] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 838.334676] env[61985]: value = "task-935867" [ 838.334676] env[61985]: _type = "Task" [ 838.334676] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.346933] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935867, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.361027] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520cd259-8f04-865b-f9a5-582f54db7aac, 'name': SearchDatastore_Task, 'duration_secs': 0.011104} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.361689] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.361987] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475/83e4ecc0-cc96-48e1-8c1d-7dcf84d72475.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 838.362283] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-86c04a08-08e9-4e6d-9061-4b1125ead90a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.371420] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 838.371420] env[61985]: value = "task-935868" [ 838.371420] env[61985]: _type = "Task" [ 838.371420] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.382976] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935868, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.565867] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 838.566226] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 838.566398] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Rebuilding the list of instances to heal {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9936}} [ 838.610656] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935866, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.710129] env[61985]: DEBUG nova.network.neutron [req-ee98f274-153c-4dd7-a029-ac47d46d60e8 req-c864b820-6b10-4a61-8892-187f539782fe service nova] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Updated VIF entry in instance network info cache for port 5b8f2f2c-0511-4436-982f-3f32c10f5759. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 838.710349] env[61985]: DEBUG nova.network.neutron [req-ee98f274-153c-4dd7-a029-ac47d46d60e8 req-c864b820-6b10-4a61-8892-187f539782fe service nova] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Updating instance_info_cache with network_info: [{"id": "5b8f2f2c-0511-4436-982f-3f32c10f5759", "address": "fa:16:3e:ff:c2:cf", "network": {"id": "ddf63d11-2633-4be8-b73f-c5ca846b420c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-709607861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8015ef21eaf4a0bb19063e261064cdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b8f2f2c-05", "ovs_interfaceid": "5b8f2f2c-0511-4436-982f-3f32c10f5759", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.797934] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.729s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.799911] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 838.802576] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.553s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.803587] env[61985]: INFO nova.compute.claims [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.854639] env[61985]: INFO nova.compute.manager [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Took 41.65 seconds to build instance. [ 838.862620] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935867, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091872} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.862620] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.863424] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d43b282-0e86-4bc4-83bd-65078104ed16 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.891020] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf/5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.894655] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f3ed9f4-94e5-457b-9b45-99d729aac187 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.916699] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935868, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504888} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.918111] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475/83e4ecc0-cc96-48e1-8c1d-7dcf84d72475.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 838.918371] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.918707] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 838.918707] env[61985]: value = "task-935869" [ 838.918707] env[61985]: _type = "Task" [ 838.918707] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.918949] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c1307a4c-81ef-4f3a-b0c8-48187ad935de {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.931904] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935869, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.932480] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 838.932480] env[61985]: value = "task-935870" [ 838.932480] env[61985]: _type = "Task" [ 838.932480] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.941764] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935870, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.073604] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 839.074848] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 839.074848] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 839.074848] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 839.074848] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.074848] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquired lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.074848] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Forcefully refreshing network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 839.075292] env[61985]: DEBUG nova.objects.instance [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lazy-loading 'info_cache' on Instance uuid 2322ab72-9841-41fb-9d60-2812baabe108 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.107254] env[61985]: DEBUG oslo_vmware.api [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935866, 'name': RemoveSnapshot_Task, 'duration_secs': 0.595348} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.107613] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 839.107908] env[61985]: INFO nova.compute.manager [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Took 16.40 seconds to snapshot the instance on the hypervisor. [ 839.214652] env[61985]: DEBUG oslo_concurrency.lockutils [req-ee98f274-153c-4dd7-a029-ac47d46d60e8 req-c864b820-6b10-4a61-8892-187f539782fe service nova] Releasing lock "refresh_cache-83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.247133] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Successfully updated port: d3b72151-261d-424c-adf2-aa8c1b98ec5e {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.308528] env[61985]: DEBUG nova.compute.utils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.311337] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 839.311337] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 839.357050] env[61985]: DEBUG oslo_concurrency.lockutils [None req-afaad702-d2dc-45c4-ad1f-3fbc5efceb8f tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "c4e95f51-9acc-46e8-9921-e0a85fb38598" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.969s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.367373] env[61985]: DEBUG nova.policy [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3d8fab40f554eb3a4804a4291686146', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c06fc8ffa004e00be0a1991d2e082e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.437077] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935869, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.445889] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935870, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076547} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.445889] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.445889] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e872b83a-b917-4890-8ca2-991a447518c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.471063] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475/83e4ecc0-cc96-48e1-8c1d-7dcf84d72475.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.471063] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d47ff027-e1aa-4416-8974-c987bfef7161 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.493917] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 839.493917] env[61985]: value = "task-935871" [ 839.493917] env[61985]: _type = "Task" [ 839.493917] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.510264] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935871, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.570125] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquiring lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.570390] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.667197] env[61985]: DEBUG nova.compute.manager [req-82bb202f-6f64-4ce9-a6fd-bca0e6555d5e req-895eabd7-01c7-4504-8d64-fd653bfa70e6 service nova] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Received event network-vif-plugged-d3b72151-261d-424c-adf2-aa8c1b98ec5e {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 839.668402] env[61985]: DEBUG oslo_concurrency.lockutils [req-82bb202f-6f64-4ce9-a6fd-bca0e6555d5e req-895eabd7-01c7-4504-8d64-fd653bfa70e6 service nova] Acquiring lock "95e3e6d7-9203-47a0-bad7-050eba09c511-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.668402] env[61985]: DEBUG oslo_concurrency.lockutils [req-82bb202f-6f64-4ce9-a6fd-bca0e6555d5e req-895eabd7-01c7-4504-8d64-fd653bfa70e6 service nova] Lock "95e3e6d7-9203-47a0-bad7-050eba09c511-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.668402] env[61985]: DEBUG oslo_concurrency.lockutils [req-82bb202f-6f64-4ce9-a6fd-bca0e6555d5e req-895eabd7-01c7-4504-8d64-fd653bfa70e6 service nova] Lock "95e3e6d7-9203-47a0-bad7-050eba09c511-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.668691] env[61985]: DEBUG nova.compute.manager [req-82bb202f-6f64-4ce9-a6fd-bca0e6555d5e req-895eabd7-01c7-4504-8d64-fd653bfa70e6 service nova] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] No waiting events found dispatching network-vif-plugged-d3b72151-261d-424c-adf2-aa8c1b98ec5e {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.668953] env[61985]: WARNING nova.compute.manager [req-82bb202f-6f64-4ce9-a6fd-bca0e6555d5e req-895eabd7-01c7-4504-8d64-fd653bfa70e6 service nova] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Received unexpected event network-vif-plugged-d3b72151-261d-424c-adf2-aa8c1b98ec5e for instance with vm_state building and task_state spawning. [ 839.687745] env[61985]: DEBUG nova.compute.manager [None req-62451aac-85b1-4cca-928d-53dd620ad3ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Found 1 images (rotation: 2) {{(pid=61985) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4560}} [ 839.750260] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "refresh_cache-95e3e6d7-9203-47a0-bad7-050eba09c511" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.750416] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "refresh_cache-95e3e6d7-9203-47a0-bad7-050eba09c511" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.750566] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 839.820597] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 839.858765] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquiring lock "402ef006-2835-4239-a3c4-1c37a7bba9a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.860279] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lock "402ef006-2835-4239-a3c4-1c37a7bba9a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.862379] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquiring lock "402ef006-2835-4239-a3c4-1c37a7bba9a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.862379] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lock "402ef006-2835-4239-a3c4-1c37a7bba9a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.862379] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lock "402ef006-2835-4239-a3c4-1c37a7bba9a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.868040] env[61985]: DEBUG nova.compute.manager [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 839.872202] env[61985]: INFO nova.compute.manager [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Terminating instance [ 839.876031] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Successfully created port: ae5604d8-2112-4be8-a6d8-d5a94b4eac44 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.878490] env[61985]: DEBUG nova.compute.manager [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 839.878777] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.880039] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d3ae0f-c716-44b1-ba84-1bcdf1bcf1a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.895475] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.895985] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41773a9d-605d-48b5-8fcf-db96f0ce07c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.906527] env[61985]: DEBUG oslo_vmware.api [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for the task: (returnval){ [ 839.906527] env[61985]: value = "task-935872" [ 839.906527] env[61985]: _type = "Task" [ 839.906527] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.923517] env[61985]: DEBUG oslo_vmware.api [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935872, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.936460] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935869, 'name': ReconfigVM_Task, 'duration_secs': 0.962925} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.937108] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf/5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.937955] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1218a86c-42bd-4b86-8f56-48bd60ab1ed7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.950460] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 839.950460] env[61985]: value = "task-935873" [ 839.950460] env[61985]: _type = "Task" [ 839.950460] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.961541] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935873, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.005459] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935871, 'name': ReconfigVM_Task, 'duration_secs': 0.315426} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.005779] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475/83e4ecc0-cc96-48e1-8c1d-7dcf84d72475.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.008986] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-03c365ba-d46c-4bbe-bdaa-02cec16d7af8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.017320] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 840.017320] env[61985]: value = "task-935874" [ 840.017320] env[61985]: _type = "Task" [ 840.017320] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.028753] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935874, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.287310] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee913ac-6ef7-4bc1-92b2-82cbbb9810e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.300138] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 840.303863] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e77269e6-9a52-4143-8c61-5dfc76a783b6 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Suspending the VM {{(pid=61985) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 840.304179] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-5834db2c-57f7-4dfa-bfef-2c186eb56022 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.312370] env[61985]: DEBUG oslo_vmware.api [None req-e77269e6-9a52-4143-8c61-5dfc76a783b6 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 840.312370] env[61985]: value = "task-935875" [ 840.312370] env[61985]: _type = "Task" [ 840.312370] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.324950] env[61985]: DEBUG oslo_vmware.api [None req-e77269e6-9a52-4143-8c61-5dfc76a783b6 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935875, 'name': SuspendVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.337017] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d23f9a-66ba-44ae-9466-c420949acf6a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.342798] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f7f1a8-f197-44f0-9ef3-ebf73d6847ca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.387292] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573696a6-e742-4661-a382-721c7aeadee3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.399788] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731a83b4-b1e4-439b-a67b-c1142da7bbe1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.405154] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.419540] env[61985]: DEBUG nova.compute.provider_tree [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.429151] env[61985]: DEBUG oslo_vmware.api [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935872, 'name': PowerOffVM_Task, 'duration_secs': 0.395055} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.429151] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 840.429151] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 840.429151] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7829733c-e32f-4a74-89cc-23d19e912efb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.460396] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935873, 'name': Rename_Task, 'duration_secs': 0.336093} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.460710] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.460973] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9987ec40-3cda-4e26-945c-18273faacff8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.468123] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 840.468123] env[61985]: value = "task-935877" [ 840.468123] env[61985]: _type = "Task" [ 840.468123] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.476801] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935877, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.528361] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935874, 'name': Rename_Task, 'duration_secs': 0.164316} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.528688] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.528956] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f55a5712-7a8b-4354-b01e-e67a3c066df4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.537331] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 840.537331] env[61985]: value = "task-935878" [ 840.537331] env[61985]: _type = "Task" [ 840.537331] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.547016] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935878, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.569754] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Updating instance_info_cache with network_info: [{"id": "d3b72151-261d-424c-adf2-aa8c1b98ec5e", "address": "fa:16:3e:25:33:b7", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3b72151-26", "ovs_interfaceid": "d3b72151-261d-424c-adf2-aa8c1b98ec5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.825313] env[61985]: DEBUG oslo_vmware.api [None req-e77269e6-9a52-4143-8c61-5dfc76a783b6 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935875, 'name': SuspendVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.838832] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 840.864332] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.864490] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.864658] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.864850] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.865018] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.871258] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.871537] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.871717] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.871976] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.872185] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.872376] env[61985]: DEBUG nova.virt.hardware [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.873288] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899e5cc6-64a4-4f24-b5fd-a95a1fc6301f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.883045] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769b61b9-1620-43f2-8444-ddb7be750bf9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.922735] env[61985]: DEBUG nova.scheduler.client.report [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.987249] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935877, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.051850] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935878, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.072945] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "refresh_cache-95e3e6d7-9203-47a0-bad7-050eba09c511" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.073872] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Instance network_info: |[{"id": "d3b72151-261d-424c-adf2-aa8c1b98ec5e", "address": "fa:16:3e:25:33:b7", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3b72151-26", "ovs_interfaceid": "d3b72151-261d-424c-adf2-aa8c1b98ec5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 841.075127] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:33:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3b72151-261d-424c-adf2-aa8c1b98ec5e', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.095034] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Creating folder: Project (1c06fc8ffa004e00be0a1991d2e082e8). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.095034] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7e6db617-79e1-4881-8cc2-b72c3ddb416d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.107636] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Created folder: Project (1c06fc8ffa004e00be0a1991d2e082e8) in parent group-v211285. [ 841.107636] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Creating folder: Instances. Parent ref: group-v211355. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.107877] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6a8f219-bcb7-48dc-9ed6-e2829a899a4b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.120321] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Created folder: Instances in parent group-v211355. [ 841.120631] env[61985]: DEBUG oslo.service.loopingcall [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.120989] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 841.121123] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a1fb1db-d8d2-42ae-819f-bdc513620603 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.144667] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.144667] env[61985]: value = "task-935881" [ 841.144667] env[61985]: _type = "Task" [ 841.144667] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.153867] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935881, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.167199] env[61985]: DEBUG nova.compute.manager [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.167564] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659ae0e7-03f0-4610-b20e-d2e936e436f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.177478] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance_info_cache with network_info: [{"id": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "address": "fa:16:3e:12:45:ef", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aa18b7b-89", "ovs_interfaceid": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.328609] env[61985]: DEBUG oslo_vmware.api [None req-e77269e6-9a52-4143-8c61-5dfc76a783b6 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935875, 'name': SuspendVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.428543] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.626s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.429156] env[61985]: DEBUG nova.compute.manager [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 841.431866] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.721s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.432081] env[61985]: DEBUG nova.objects.instance [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61985) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 841.480951] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935877, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.548092] env[61985]: DEBUG oslo_vmware.api [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935878, 'name': PowerOnVM_Task, 'duration_secs': 0.60076} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.548422] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.548674] env[61985]: INFO nova.compute.manager [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Took 7.09 seconds to spawn the instance on the hypervisor. [ 841.548935] env[61985]: DEBUG nova.compute.manager [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.549633] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132eb40b-30d2-4f90-9207-fae0b610c39f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.655879] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935881, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.680529] env[61985]: INFO nova.compute.manager [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] instance snapshotting [ 841.681179] env[61985]: DEBUG nova.objects.instance [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'flavor' on Instance uuid 2322ab72-9841-41fb-9d60-2812baabe108 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.682652] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Releasing lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.682878] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updated the network info_cache for instance {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10003}} [ 841.683537] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.684076] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.684368] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.684579] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.684781] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.684917] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 841.685076] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.828714] env[61985]: DEBUG oslo_vmware.api [None req-e77269e6-9a52-4143-8c61-5dfc76a783b6 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935875, 'name': SuspendVM_Task, 'duration_secs': 1.026952} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.829487] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e77269e6-9a52-4143-8c61-5dfc76a783b6 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Suspended the VM {{(pid=61985) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 841.829487] env[61985]: DEBUG nova.compute.manager [None req-e77269e6-9a52-4143-8c61-5dfc76a783b6 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.829883] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033ecda5-85f8-45fa-9105-6c5c13203837 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.876574] env[61985]: DEBUG nova.compute.manager [req-e6395848-b495-435d-a8a0-ae0e1514158b req-ca781233-c646-4f51-ad00-73673337bd8f service nova] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Received event network-changed-d3b72151-261d-424c-adf2-aa8c1b98ec5e {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 841.876886] env[61985]: DEBUG nova.compute.manager [req-e6395848-b495-435d-a8a0-ae0e1514158b req-ca781233-c646-4f51-ad00-73673337bd8f service nova] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Refreshing instance network info cache due to event network-changed-d3b72151-261d-424c-adf2-aa8c1b98ec5e. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 841.877203] env[61985]: DEBUG oslo_concurrency.lockutils [req-e6395848-b495-435d-a8a0-ae0e1514158b req-ca781233-c646-4f51-ad00-73673337bd8f service nova] Acquiring lock "refresh_cache-95e3e6d7-9203-47a0-bad7-050eba09c511" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.877389] env[61985]: DEBUG oslo_concurrency.lockutils [req-e6395848-b495-435d-a8a0-ae0e1514158b req-ca781233-c646-4f51-ad00-73673337bd8f service nova] Acquired lock "refresh_cache-95e3e6d7-9203-47a0-bad7-050eba09c511" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.877515] env[61985]: DEBUG nova.network.neutron [req-e6395848-b495-435d-a8a0-ae0e1514158b req-ca781233-c646-4f51-ad00-73673337bd8f service nova] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Refreshing network info cache for port d3b72151-261d-424c-adf2-aa8c1b98ec5e {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 841.940510] env[61985]: DEBUG nova.compute.utils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.942359] env[61985]: DEBUG nova.compute.manager [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Not allocating networking since 'none' was specified. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 841.986022] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935877, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.001705] env[61985]: DEBUG nova.compute.manager [req-fcb504f8-97ec-42b3-8a92-bf357a07d7a0 req-04559913-ff4b-4313-922d-0783f18717b1 service nova] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Received event network-vif-plugged-ae5604d8-2112-4be8-a6d8-d5a94b4eac44 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 842.002025] env[61985]: DEBUG oslo_concurrency.lockutils [req-fcb504f8-97ec-42b3-8a92-bf357a07d7a0 req-04559913-ff4b-4313-922d-0783f18717b1 service nova] Acquiring lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.002177] env[61985]: DEBUG oslo_concurrency.lockutils [req-fcb504f8-97ec-42b3-8a92-bf357a07d7a0 req-04559913-ff4b-4313-922d-0783f18717b1 service nova] Lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.002321] env[61985]: DEBUG oslo_concurrency.lockutils [req-fcb504f8-97ec-42b3-8a92-bf357a07d7a0 req-04559913-ff4b-4313-922d-0783f18717b1 service nova] Lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.002486] env[61985]: DEBUG nova.compute.manager [req-fcb504f8-97ec-42b3-8a92-bf357a07d7a0 req-04559913-ff4b-4313-922d-0783f18717b1 service nova] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] No waiting events found dispatching network-vif-plugged-ae5604d8-2112-4be8-a6d8-d5a94b4eac44 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 842.002653] env[61985]: WARNING nova.compute.manager [req-fcb504f8-97ec-42b3-8a92-bf357a07d7a0 req-04559913-ff4b-4313-922d-0783f18717b1 service nova] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Received unexpected event network-vif-plugged-ae5604d8-2112-4be8-a6d8-d5a94b4eac44 for instance with vm_state building and task_state spawning. [ 842.071231] env[61985]: INFO nova.compute.manager [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Took 38.29 seconds to build instance. [ 842.156944] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935881, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.187838] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.188825] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c5ad924-f9db-43a7-93c8-d3f49b0c056f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.212533] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3396a7-3f8d-446f-b342-8c79932099e8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.367023] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 842.367023] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 842.367023] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Deleting the datastore file [datastore2] 402ef006-2835-4239-a3c4-1c37a7bba9a5 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 842.367023] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a4b3a60-044e-403e-9f5a-6f22e5fb58b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.375874] env[61985]: DEBUG oslo_vmware.api [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for the task: (returnval){ [ 842.375874] env[61985]: value = "task-935882" [ 842.375874] env[61985]: _type = "Task" [ 842.375874] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.377367] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Successfully updated port: ae5604d8-2112-4be8-a6d8-d5a94b4eac44 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.392370] env[61985]: DEBUG oslo_vmware.api [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.445174] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1919c915-94fc-4173-bef2-3d742ab2869c tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.445174] env[61985]: DEBUG nova.compute.manager [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 842.447971] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.487s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.448229] env[61985]: DEBUG nova.objects.instance [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lazy-loading 'resources' on Instance uuid b659ff28-5101-4825-84ea-111351c81145 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.483764] env[61985]: DEBUG oslo_vmware.api [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-935877, 'name': PowerOnVM_Task, 'duration_secs': 1.565588} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.484135] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 842.484430] env[61985]: INFO nova.compute.manager [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Took 10.55 seconds to spawn the instance on the hypervisor. [ 842.486198] env[61985]: DEBUG nova.compute.manager [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.486198] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b078a022-ebd9-4f56-b95f-672ca63c6140 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.522789] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "6dd8218b-c14d-40c9-87df-097fab06c669" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.523062] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.573443] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7b6bed42-81f7-404c-8b72-6dca14cc8fe7 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.912s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.657811] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935881, 'name': CreateVM_Task, 'duration_secs': 1.480738} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.658118] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 842.658896] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.659163] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.659872] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.659872] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0477f4a2-82c8-439a-a4a9-9013dc886c66 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.664894] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 842.664894] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b45788-35dd-2492-e99e-84ec0fc76cc3" [ 842.664894] env[61985]: _type = "Task" [ 842.664894] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.673573] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b45788-35dd-2492-e99e-84ec0fc76cc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.724426] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 842.724679] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-25bfefb7-f6eb-477c-ad8f-44789ffa2e13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.733900] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 842.733900] env[61985]: value = "task-935883" [ 842.733900] env[61985]: _type = "Task" [ 842.733900] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.743335] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935883, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.806722] env[61985]: DEBUG nova.network.neutron [req-e6395848-b495-435d-a8a0-ae0e1514158b req-ca781233-c646-4f51-ad00-73673337bd8f service nova] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Updated VIF entry in instance network info cache for port d3b72151-261d-424c-adf2-aa8c1b98ec5e. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 842.807220] env[61985]: DEBUG nova.network.neutron [req-e6395848-b495-435d-a8a0-ae0e1514158b req-ca781233-c646-4f51-ad00-73673337bd8f service nova] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Updating instance_info_cache with network_info: [{"id": "d3b72151-261d-424c-adf2-aa8c1b98ec5e", "address": "fa:16:3e:25:33:b7", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3b72151-26", "ovs_interfaceid": "d3b72151-261d-424c-adf2-aa8c1b98ec5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.883334] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "refresh_cache-e82d26b1-5502-4fd2-89c5-ffb0c1557c79" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.883492] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "refresh_cache-e82d26b1-5502-4fd2-89c5-ffb0c1557c79" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.883649] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 842.890264] env[61985]: DEBUG oslo_vmware.api [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Task: {'id': task-935882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.244274} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.891926] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.891926] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 842.891926] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 842.891926] env[61985]: INFO nova.compute.manager [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Took 3.01 seconds to destroy the instance on the hypervisor. [ 842.891926] env[61985]: DEBUG oslo.service.loopingcall [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.892423] env[61985]: DEBUG nova.compute.manager [-] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 842.892698] env[61985]: DEBUG nova.network.neutron [-] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 843.015671] env[61985]: INFO nova.compute.manager [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Took 41.06 seconds to build instance. [ 843.078714] env[61985]: DEBUG nova.compute.manager [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.183977] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b45788-35dd-2492-e99e-84ec0fc76cc3, 'name': SearchDatastore_Task, 'duration_secs': 0.02656} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.187201] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.187496] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.187792] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.187971] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.188195] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.188497] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed9312ef-69d1-4a9d-8b8f-c386cc4ffd2d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.204672] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.205182] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 843.206970] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c9e038e-4e6a-4986-b0c1-a39dd72cce19 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.217045] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 843.217045] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52bb93d7-8c12-8111-ab30-abfeb42d880b" [ 843.217045] env[61985]: _type = "Task" [ 843.217045] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.227054] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52bb93d7-8c12-8111-ab30-abfeb42d880b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.247330] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935883, 'name': CreateSnapshot_Task, 'duration_secs': 0.467786} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.247633] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 843.248501] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41bc67a-2d66-4a81-a4ff-c94b3ada3ee7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.316172] env[61985]: DEBUG oslo_concurrency.lockutils [req-e6395848-b495-435d-a8a0-ae0e1514158b req-ca781233-c646-4f51-ad00-73673337bd8f service nova] Releasing lock "refresh_cache-95e3e6d7-9203-47a0-bad7-050eba09c511" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.442531] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 843.469195] env[61985]: DEBUG nova.compute.manager [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 843.493439] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.493712] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.493847] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.494053] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.494235] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.494354] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.494572] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.494955] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.495076] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.495140] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.495384] env[61985]: DEBUG nova.virt.hardware [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.496239] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d227cff-3e7e-4df1-bd5d-a4b9befb925c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.507849] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8445565e-86b7-4b3a-b36d-dcd15208e1b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.522416] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb9eb342-2e02-4e29-a637-7ac83d1314b7 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.798s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.522898] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.528406] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Creating folder: Project (fc5f77e8b8ff4e458c26d899ebbb60e4). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 843.532159] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9f91a4e-5e80-43d4-9183-511f66b4e9c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.547033] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Created folder: Project (fc5f77e8b8ff4e458c26d899ebbb60e4) in parent group-v211285. [ 843.547324] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Creating folder: Instances. Parent ref: group-v211359. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 843.547562] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20dbd53f-0169-45e0-97ec-df611ec0d3d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.554905] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-307f0cad-a488-43dc-987f-e8cb5e5fd23f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.564272] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d16503b-833e-4518-a199-3307469f7c22 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.567420] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Created folder: Instances in parent group-v211359. [ 843.567698] env[61985]: DEBUG oslo.service.loopingcall [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.567906] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 843.568504] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-93682cfc-f836-4858-af6e-e1065f63b02e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.616176] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0bfc1c-3439-4200-b692-ced48b7e85b6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.622010] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.622010] env[61985]: value = "task-935886" [ 843.622010] env[61985]: _type = "Task" [ 843.622010] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.629194] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b91242-b314-4d29-adb0-e9462043d7e2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.636285] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935886, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.637332] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.650361] env[61985]: DEBUG nova.compute.provider_tree [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.729283] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52bb93d7-8c12-8111-ab30-abfeb42d880b, 'name': SearchDatastore_Task, 'duration_secs': 0.032935} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.730201] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7a7fa11-e030-43d0-9aa9-71962ee21e28 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.737012] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 843.737012] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5245b505-4284-a9cf-b920-2f92170c6cd3" [ 843.737012] env[61985]: _type = "Task" [ 843.737012] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.750583] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5245b505-4284-a9cf-b920-2f92170c6cd3, 'name': SearchDatastore_Task, 'duration_secs': 0.011383} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.751133] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.751644] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 95e3e6d7-9203-47a0-bad7-050eba09c511/95e3e6d7-9203-47a0-bad7-050eba09c511.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 843.751774] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b01ca2bd-cbe1-490d-98fc-1238bc3b6718 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.754971] env[61985]: DEBUG nova.network.neutron [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Updating instance_info_cache with network_info: [{"id": "ae5604d8-2112-4be8-a6d8-d5a94b4eac44", "address": "fa:16:3e:36:83:ee", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae5604d8-21", "ovs_interfaceid": "ae5604d8-2112-4be8-a6d8-d5a94b4eac44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.763512] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 843.763512] env[61985]: value = "task-935887" [ 843.763512] env[61985]: _type = "Task" [ 843.763512] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.774509] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 843.775353] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4c309d93-083f-4f11-8e6b-a11108459697 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.786897] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935887, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.788497] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 843.788497] env[61985]: value = "task-935888" [ 843.788497] env[61985]: _type = "Task" [ 843.788497] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.798358] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.865142] env[61985]: DEBUG nova.network.neutron [-] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.037422] env[61985]: DEBUG nova.compute.manager [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 844.136263] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935886, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.153505] env[61985]: DEBUG nova.scheduler.client.report [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.217097] env[61985]: DEBUG nova.compute.manager [req-2941b194-7c4a-4834-84f7-6dfb5379ee83 req-b6b5129c-e1ee-4811-9223-2f1c7cf6c489 service nova] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Received event network-changed-ae5604d8-2112-4be8-a6d8-d5a94b4eac44 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 844.217811] env[61985]: DEBUG nova.compute.manager [req-2941b194-7c4a-4834-84f7-6dfb5379ee83 req-b6b5129c-e1ee-4811-9223-2f1c7cf6c489 service nova] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Refreshing instance network info cache due to event network-changed-ae5604d8-2112-4be8-a6d8-d5a94b4eac44. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 844.218028] env[61985]: DEBUG oslo_concurrency.lockutils [req-2941b194-7c4a-4834-84f7-6dfb5379ee83 req-b6b5129c-e1ee-4811-9223-2f1c7cf6c489 service nova] Acquiring lock "refresh_cache-e82d26b1-5502-4fd2-89c5-ffb0c1557c79" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.258153] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "refresh_cache-e82d26b1-5502-4fd2-89c5-ffb0c1557c79" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.258480] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Instance network_info: |[{"id": "ae5604d8-2112-4be8-a6d8-d5a94b4eac44", "address": "fa:16:3e:36:83:ee", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae5604d8-21", "ovs_interfaceid": "ae5604d8-2112-4be8-a6d8-d5a94b4eac44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.258805] env[61985]: DEBUG oslo_concurrency.lockutils [req-2941b194-7c4a-4834-84f7-6dfb5379ee83 req-b6b5129c-e1ee-4811-9223-2f1c7cf6c489 service nova] Acquired lock "refresh_cache-e82d26b1-5502-4fd2-89c5-ffb0c1557c79" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.258996] env[61985]: DEBUG nova.network.neutron [req-2941b194-7c4a-4834-84f7-6dfb5379ee83 req-b6b5129c-e1ee-4811-9223-2f1c7cf6c489 service nova] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Refreshing network info cache for port ae5604d8-2112-4be8-a6d8-d5a94b4eac44 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 844.263178] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:83:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ae5604d8-2112-4be8-a6d8-d5a94b4eac44', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.268560] env[61985]: DEBUG oslo.service.loopingcall [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.269154] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 844.273359] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-50e62e13-96e8-4d2a-ae4f-01ec7b9cb7de {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.298127] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935887, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.299784] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.299784] env[61985]: value = "task-935889" [ 844.299784] env[61985]: _type = "Task" [ 844.299784] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.306993] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task} progress is 93%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.314120] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935889, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.371810] env[61985]: INFO nova.compute.manager [-] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Took 1.48 seconds to deallocate network for instance. [ 844.563194] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.637260] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935886, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.659474] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.211s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.662617] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.332s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.662617] env[61985]: DEBUG nova.objects.instance [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lazy-loading 'resources' on Instance uuid 4aca0ecb-4ae6-4400-accd-d71782b2806d {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.689113] env[61985]: INFO nova.scheduler.client.report [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Deleted allocations for instance b659ff28-5101-4825-84ea-111351c81145 [ 844.757171] env[61985]: DEBUG nova.compute.manager [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.758447] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ddb22b5-4bc5-4dad-b306-ee53d582716d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.780631] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935887, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606859} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.781315] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 95e3e6d7-9203-47a0-bad7-050eba09c511/95e3e6d7-9203-47a0-bad7-050eba09c511.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 844.781651] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.781975] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7314673-eda5-49b7-bc43-eed0a4b5c84a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.795434] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 844.795434] env[61985]: value = "task-935890" [ 844.795434] env[61985]: _type = "Task" [ 844.795434] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.802862] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task} progress is 93%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.811044] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.817064] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935889, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.879617] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.139621] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935886, 'name': CreateVM_Task, 'duration_secs': 1.451208} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.140167] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 845.140714] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.140930] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.141437] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.141749] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-575654ca-d20c-4d9a-817f-fe9a2e82bd55 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.147614] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for the task: (returnval){ [ 845.147614] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c23e95-5301-3131-f212-2b531f364e61" [ 845.147614] env[61985]: _type = "Task" [ 845.147614] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.157064] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c23e95-5301-3131-f212-2b531f364e61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.196017] env[61985]: DEBUG nova.network.neutron [req-2941b194-7c4a-4834-84f7-6dfb5379ee83 req-b6b5129c-e1ee-4811-9223-2f1c7cf6c489 service nova] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Updated VIF entry in instance network info cache for port ae5604d8-2112-4be8-a6d8-d5a94b4eac44. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 845.196017] env[61985]: DEBUG nova.network.neutron [req-2941b194-7c4a-4834-84f7-6dfb5379ee83 req-b6b5129c-e1ee-4811-9223-2f1c7cf6c489 service nova] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Updating instance_info_cache with network_info: [{"id": "ae5604d8-2112-4be8-a6d8-d5a94b4eac44", "address": "fa:16:3e:36:83:ee", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae5604d8-21", "ovs_interfaceid": "ae5604d8-2112-4be8-a6d8-d5a94b4eac44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.203621] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7a3eff04-0e6b-4bfe-987c-c2cf99e91cb9 tempest-SecurityGroupsTestJSON-1099982380 tempest-SecurityGroupsTestJSON-1099982380-project-member] Lock "b659ff28-5101-4825-84ea-111351c81145" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.378s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.270370] env[61985]: INFO nova.compute.manager [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] instance snapshotting [ 845.270649] env[61985]: WARNING nova.compute.manager [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 845.275802] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e15ec4b-a795-4965-b6d9-b42ce2aa4c0e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.301413] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6db547d-e69f-4dac-b019-c44cf1f908fe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.314260] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.326869] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110358} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.327050] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935889, 'name': CreateVM_Task, 'duration_secs': 0.780659} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.327751] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.327970] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 845.328727] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b6f6d9-8cb7-43c7-a311-7a5ad2a9b041 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.331568] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.352734] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 95e3e6d7-9203-47a0-bad7-050eba09c511/95e3e6d7-9203-47a0-bad7-050eba09c511.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.356049] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a7db965-06c2-4900-8a38-98f0d567594e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.379177] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 845.379177] env[61985]: value = "task-935891" [ 845.379177] env[61985]: _type = "Task" [ 845.379177] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.392851] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935891, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.639246] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd4e04e-4123-4a5b-b0c8-974df3a610b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.649040] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae0d5bb-3bbc-40f1-a624-1da3c8315b91 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.663409] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c23e95-5301-3131-f212-2b531f364e61, 'name': SearchDatastore_Task, 'duration_secs': 0.022946} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.690669] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.690928] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.691881] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.691881] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.691881] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.692190] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.693605] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.693605] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0db2547-dd85-4501-aa26-776c8085fab3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.695836] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73fed1d-130f-40cd-afe9-c99b00be68eb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.698572] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ce63307-9bac-4a47-ab68-32d8f2956bde {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.702120] env[61985]: DEBUG oslo_concurrency.lockutils [req-2941b194-7c4a-4834-84f7-6dfb5379ee83 req-b6b5129c-e1ee-4811-9223-2f1c7cf6c489 service nova] Releasing lock "refresh_cache-e82d26b1-5502-4fd2-89c5-ffb0c1557c79" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.702365] env[61985]: DEBUG nova.compute.manager [req-2941b194-7c4a-4834-84f7-6dfb5379ee83 req-b6b5129c-e1ee-4811-9223-2f1c7cf6c489 service nova] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Received event network-vif-deleted-c32c52fa-48c3-48d7-87e9-0e9d693d6959 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 845.710204] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eabebc93-7dd3-40eb-95f5-a72d0389eacb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.715087] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 845.715087] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521eea86-6750-8097-e50e-634919294580" [ 845.715087] env[61985]: _type = "Task" [ 845.715087] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.716280] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.716461] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 845.720251] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-369e683d-add0-40c8-8574-659d1517ee8c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.731664] env[61985]: DEBUG nova.compute.provider_tree [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.739739] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521eea86-6750-8097-e50e-634919294580, 'name': SearchDatastore_Task, 'duration_secs': 0.019899} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.741357] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.741607] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.741826] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.742121] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for the task: (returnval){ [ 845.742121] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a50d9c-e3e4-5c8e-7c1c-87c0bd696a4c" [ 845.742121] env[61985]: _type = "Task" [ 845.742121] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.751412] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a50d9c-e3e4-5c8e-7c1c-87c0bd696a4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.805647] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.829576] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 845.829856] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b924cccc-bbd8-4169-9642-91ca07fbd214 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.840163] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 845.840163] env[61985]: value = "task-935892" [ 845.840163] env[61985]: _type = "Task" [ 845.840163] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.849512] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935892, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.890237] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935891, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.236517] env[61985]: DEBUG nova.scheduler.client.report [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.255299] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a50d9c-e3e4-5c8e-7c1c-87c0bd696a4c, 'name': SearchDatastore_Task, 'duration_secs': 0.019742} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.256646] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b4214c2-c27f-46c0-94ac-95ff8b1fabf0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.263304] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for the task: (returnval){ [ 846.263304] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529ccab6-7128-dead-a1ff-a29e9e8e5c34" [ 846.263304] env[61985]: _type = "Task" [ 846.263304] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.273830] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529ccab6-7128-dead-a1ff-a29e9e8e5c34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.307873] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.350688] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935892, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.353942] env[61985]: DEBUG nova.compute.manager [req-fd572b2a-acad-4a8d-b691-250b5403d24d req-c76a35bd-4e84-4692-8511-fabc604b7e14 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received event network-changed-4959d778-7709-41bf-a558-f6ba7cce7d4b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 846.353942] env[61985]: DEBUG nova.compute.manager [req-fd572b2a-acad-4a8d-b691-250b5403d24d req-c76a35bd-4e84-4692-8511-fabc604b7e14 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Refreshing instance network info cache due to event network-changed-4959d778-7709-41bf-a558-f6ba7cce7d4b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 846.353942] env[61985]: DEBUG oslo_concurrency.lockutils [req-fd572b2a-acad-4a8d-b691-250b5403d24d req-c76a35bd-4e84-4692-8511-fabc604b7e14 service nova] Acquiring lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.354302] env[61985]: DEBUG oslo_concurrency.lockutils [req-fd572b2a-acad-4a8d-b691-250b5403d24d req-c76a35bd-4e84-4692-8511-fabc604b7e14 service nova] Acquired lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.354405] env[61985]: DEBUG nova.network.neutron [req-fd572b2a-acad-4a8d-b691-250b5403d24d req-c76a35bd-4e84-4692-8511-fabc604b7e14 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Refreshing network info cache for port 4959d778-7709-41bf-a558-f6ba7cce7d4b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 846.390721] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935891, 'name': ReconfigVM_Task, 'duration_secs': 0.85766} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.391060] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 95e3e6d7-9203-47a0-bad7-050eba09c511/95e3e6d7-9203-47a0-bad7-050eba09c511.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.391648] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8da925a9-7e28-43b7-805b-5446f4eba68e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.400030] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 846.400030] env[61985]: value = "task-935893" [ 846.400030] env[61985]: _type = "Task" [ 846.400030] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.412683] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935893, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.741508] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.079s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.743856] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.880s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.744119] env[61985]: DEBUG nova.objects.instance [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lazy-loading 'resources' on Instance uuid 794cff68-6e26-4607-96f7-eaeb41182551 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.769696] env[61985]: INFO nova.scheduler.client.report [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Deleted allocations for instance 4aca0ecb-4ae6-4400-accd-d71782b2806d [ 846.780557] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529ccab6-7128-dead-a1ff-a29e9e8e5c34, 'name': SearchDatastore_Task, 'duration_secs': 0.011387} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.780855] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.781135] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] f8924f18-7232-4be6-84bf-1ba05b855ffe/f8924f18-7232-4be6-84bf-1ba05b855ffe.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 846.781442] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.782246] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.782246] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9b6515a-5332-4e61-a9e2-a3cba8ae969d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.785030] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8152e1d1-482b-4788-860e-418e4d03690b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.794801] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for the task: (returnval){ [ 846.794801] env[61985]: value = "task-935894" [ 846.794801] env[61985]: _type = "Task" [ 846.794801] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.796171] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.796450] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 846.803954] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98aa67de-f037-4153-b160-5aac2f04e48c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.815604] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.815740] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.816729] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 846.816729] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520d95aa-d67b-e5b2-a62a-1db74de0b432" [ 846.816729] env[61985]: _type = "Task" [ 846.816729] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.828224] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520d95aa-d67b-e5b2-a62a-1db74de0b432, 'name': SearchDatastore_Task, 'duration_secs': 0.01017} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.829075] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23e4a52c-57cc-4d61-b7b6-519e371399ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.834379] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 846.834379] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529dcd66-f90f-74b3-f4ca-798febcb7c2a" [ 846.834379] env[61985]: _type = "Task" [ 846.834379] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.843935] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529dcd66-f90f-74b3-f4ca-798febcb7c2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.852088] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935892, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.910582] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935893, 'name': Rename_Task, 'duration_secs': 0.300323} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.910871] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 846.911137] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70048ac9-fef7-4ada-903c-a1cb82d4064f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.919109] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 846.919109] env[61985]: value = "task-935895" [ 846.919109] env[61985]: _type = "Task" [ 846.919109] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.927827] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935895, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.179400] env[61985]: DEBUG nova.compute.manager [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.180373] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d94858a-ee29-4069-819e-6ea4ddd7e860 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.278471] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4d717301-c1ec-4884-a3a2-f317d6aec5aa tempest-InstanceActionsNegativeTestJSON-537547174 tempest-InstanceActionsNegativeTestJSON-537547174-project-member] Lock "4aca0ecb-4ae6-4400-accd-d71782b2806d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.091s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.313289] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.317604] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935894, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.344954] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529dcd66-f90f-74b3-f4ca-798febcb7c2a, 'name': SearchDatastore_Task, 'duration_secs': 0.00947} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.348785] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.349062] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] e82d26b1-5502-4fd2-89c5-ffb0c1557c79/e82d26b1-5502-4fd2-89c5-ffb0c1557c79.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 847.352374] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-74a36506-3905-4a76-9b17-558adbd6951e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.361429] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935892, 'name': CreateSnapshot_Task, 'duration_secs': 1.137804} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.362723] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 847.363134] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 847.363134] env[61985]: value = "task-935896" [ 847.363134] env[61985]: _type = "Task" [ 847.363134] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.364116] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bc9a50-24c6-4593-959b-f8b87ea4a40a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.386652] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935896, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.432984] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935895, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.532209] env[61985]: DEBUG nova.network.neutron [req-fd572b2a-acad-4a8d-b691-250b5403d24d req-c76a35bd-4e84-4692-8511-fabc604b7e14 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updated VIF entry in instance network info cache for port 4959d778-7709-41bf-a558-f6ba7cce7d4b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 847.532300] env[61985]: DEBUG nova.network.neutron [req-fd572b2a-acad-4a8d-b691-250b5403d24d req-c76a35bd-4e84-4692-8511-fabc604b7e14 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updating instance_info_cache with network_info: [{"id": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "address": "fa:16:3e:68:94:dd", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4959d778-77", "ovs_interfaceid": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.694649] env[61985]: INFO nova.compute.manager [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] instance snapshotting [ 847.700351] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037ecf82-ac9f-4fc7-96f9-6a5bbc4947e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.728418] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e92b6ad-7c2d-4750-9d96-bd1304ef68c9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.731756] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c6e0426-fbbf-469b-8370-dddea3affafe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.742692] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a3887fc-ba3b-4416-88e9-843e20ad49cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.777317] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18fc969b-4036-4a9b-95e1-8fdfcdd5ac40 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.785330] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efed00d0-0665-4b49-b8d2-b5f411eb613f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.803408] env[61985]: DEBUG nova.compute.provider_tree [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.819418] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935894, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526083} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.822796] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] f8924f18-7232-4be6-84bf-1ba05b855ffe/f8924f18-7232-4be6-84bf-1ba05b855ffe.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 847.822796] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.822942] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.823160] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-164aaa3b-1329-41ed-b36f-ee1c2b536066 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.831023] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for the task: (returnval){ [ 847.831023] env[61985]: value = "task-935897" [ 847.831023] env[61985]: _type = "Task" [ 847.831023] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.841040] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.880500] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935896, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.900108] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 847.900108] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d8eb170e-1c77-4abb-8f17-1a6ce76cc029 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.907548] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 847.907548] env[61985]: value = "task-935898" [ 847.907548] env[61985]: _type = "Task" [ 847.907548] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.917939] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935898, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.928059] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935895, 'name': PowerOnVM_Task, 'duration_secs': 0.744912} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.928346] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 847.928598] env[61985]: INFO nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Took 10.85 seconds to spawn the instance on the hypervisor. [ 847.928839] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.929678] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873cfc6c-8dbb-42d9-97f9-eb57d3d53429 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.035427] env[61985]: DEBUG oslo_concurrency.lockutils [req-fd572b2a-acad-4a8d-b691-250b5403d24d req-c76a35bd-4e84-4692-8511-fabc604b7e14 service nova] Releasing lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.251909] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 848.252322] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-af837a10-88c3-45fd-b551-ebf47e9dabc9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.261940] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 848.261940] env[61985]: value = "task-935899" [ 848.261940] env[61985]: _type = "Task" [ 848.261940] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.271433] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935899, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.314245] env[61985]: DEBUG nova.scheduler.client.report [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.318733] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.342489] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063393} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.342771] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.345779] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8059f3c-9cff-4f92-90bf-43349bfef51f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.366094] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] f8924f18-7232-4be6-84bf-1ba05b855ffe/f8924f18-7232-4be6-84bf-1ba05b855ffe.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.366804] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1799cff5-1c08-44b3-ac19-b1263ae893cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.392419] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935896, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.924774} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.393240] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] e82d26b1-5502-4fd2-89c5-ffb0c1557c79/e82d26b1-5502-4fd2-89c5-ffb0c1557c79.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 848.393612] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.394584] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f8d0d8b9-06b0-483f-8791-830c273889fe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.396815] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for the task: (returnval){ [ 848.396815] env[61985]: value = "task-935900" [ 848.396815] env[61985]: _type = "Task" [ 848.396815] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.407462] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935900, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.409989] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 848.409989] env[61985]: value = "task-935901" [ 848.409989] env[61985]: _type = "Task" [ 848.409989] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.426016] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935901, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.429226] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935898, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.469021] env[61985]: INFO nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Took 40.52 seconds to build instance. [ 848.780083] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935899, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.820192] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.076s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.822470] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935888, 'name': CloneVM_Task, 'duration_secs': 4.778759} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.823042] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.711s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.823284] env[61985]: DEBUG nova.objects.instance [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lazy-loading 'resources' on Instance uuid f8bc8b71-0317-479d-b2f9-9471757f0774 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.824563] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Created linked-clone VM from snapshot [ 848.825650] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02d9560-57af-4be0-b154-cf85ae437d2b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.839176] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Uploading image 288950a9-1d99-44ba-aa92-2841a4ca12ea {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 848.867965] env[61985]: INFO nova.scheduler.client.report [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Deleted allocations for instance 794cff68-6e26-4607-96f7-eaeb41182551 [ 848.874452] env[61985]: DEBUG oslo_vmware.rw_handles [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 848.874452] env[61985]: value = "vm-211362" [ 848.874452] env[61985]: _type = "VirtualMachine" [ 848.874452] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 848.874662] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b3b9cfd2-c76a-4f61-8087-5fe37c285061 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.884515] env[61985]: DEBUG oslo_vmware.rw_handles [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lease: (returnval){ [ 848.884515] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5297211a-2f28-df4c-a4d9-9031145002a4" [ 848.884515] env[61985]: _type = "HttpNfcLease" [ 848.884515] env[61985]: } obtained for exporting VM: (result){ [ 848.884515] env[61985]: value = "vm-211362" [ 848.884515] env[61985]: _type = "VirtualMachine" [ 848.884515] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 848.884827] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the lease: (returnval){ [ 848.884827] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5297211a-2f28-df4c-a4d9-9031145002a4" [ 848.884827] env[61985]: _type = "HttpNfcLease" [ 848.884827] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 848.891870] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 848.891870] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5297211a-2f28-df4c-a4d9-9031145002a4" [ 848.891870] env[61985]: _type = "HttpNfcLease" [ 848.891870] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 848.906179] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935900, 'name': ReconfigVM_Task, 'duration_secs': 0.330599} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.906484] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Reconfigured VM instance instance-00000039 to attach disk [datastore2] f8924f18-7232-4be6-84bf-1ba05b855ffe/f8924f18-7232-4be6-84bf-1ba05b855ffe.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.907212] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-909d689f-4925-4590-ad6f-a4fed6dd0680 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.922376] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for the task: (returnval){ [ 848.922376] env[61985]: value = "task-935903" [ 848.922376] env[61985]: _type = "Task" [ 848.922376] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.922742] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935898, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.933188] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935901, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077529} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.937267] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.938393] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935903, 'name': Rename_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.939031] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7210ad3f-5329-4dab-a84d-2eb040cbc230 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.968067] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] e82d26b1-5502-4fd2-89c5-ffb0c1557c79/e82d26b1-5502-4fd2-89c5-ffb0c1557c79.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.968178] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c184652-b5d3-4f58-aa99-f11c8b30e6f5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.983764] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "95e3e6d7-9203-47a0-bad7-050eba09c511" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.747s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.991298] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 848.991298] env[61985]: value = "task-935904" [ 848.991298] env[61985]: _type = "Task" [ 848.991298] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.001839] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935904, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.272867] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935899, 'name': CreateSnapshot_Task, 'duration_secs': 0.999359} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.273453] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 849.274351] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14155c02-277e-4207-b481-d11d93233cb8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.382518] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9e22b9e-2a07-4103-9a7b-c873e2c01f54 tempest-ServersAdmin275Test-935008094 tempest-ServersAdmin275Test-935008094-project-member] Lock "794cff68-6e26-4607-96f7-eaeb41182551" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.363s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.404083] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.404083] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5297211a-2f28-df4c-a4d9-9031145002a4" [ 849.404083] env[61985]: _type = "HttpNfcLease" [ 849.404083] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 849.404922] env[61985]: DEBUG oslo_vmware.rw_handles [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 849.404922] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5297211a-2f28-df4c-a4d9-9031145002a4" [ 849.404922] env[61985]: _type = "HttpNfcLease" [ 849.404922] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 849.405590] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e29f70-ce65-4985-9ad7-9e0ed7e0ba40 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.422773] env[61985]: DEBUG oslo_vmware.rw_handles [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5239c6fe-8ff3-369f-c621-8095d63dff1e/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 849.423432] env[61985]: DEBUG oslo_vmware.rw_handles [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5239c6fe-8ff3-369f-c621-8095d63dff1e/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 849.430983] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935898, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.500924] env[61985]: DEBUG nova.compute.manager [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 849.518099] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935903, 'name': Rename_Task, 'duration_secs': 0.162352} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.518099] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 849.518368] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f9e913b-d327-44ef-b6ec-0090dab83d46 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.523975] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935904, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.529416] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for the task: (returnval){ [ 849.529416] env[61985]: value = "task-935905" [ 849.529416] env[61985]: _type = "Task" [ 849.529416] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.538954] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935905, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.557867] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7615a743-4db6-4f7b-886d-726d70aa162a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.793376] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 849.793833] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-16e235d4-f1b2-42b7-b351-1c45cda37d90 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.805264] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 849.805264] env[61985]: value = "task-935906" [ 849.805264] env[61985]: _type = "Task" [ 849.805264] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.814800] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935906, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.924292] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935898, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.949454] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4efbbe0c-7663-4dcd-b26b-104ac622e56f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.959011] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e81dd2f-db45-46ea-8af9-a1a54139c077 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.994707] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405730a1-852c-4019-9e03-aa1ca3af6d5b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.007583] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6c4ff4-6b31-4741-aa50-0f5d2676a066 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.022250] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935904, 'name': ReconfigVM_Task, 'duration_secs': 0.535989} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.036564] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Reconfigured VM instance instance-00000038 to attach disk [datastore2] e82d26b1-5502-4fd2-89c5-ffb0c1557c79/e82d26b1-5502-4fd2-89c5-ffb0c1557c79.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.037811] env[61985]: DEBUG nova.compute.provider_tree [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.044147] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c391564-e565-40c7-ab91-3e096865c1bf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.060269] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935905, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.064113] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.066689] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 850.066689] env[61985]: value = "task-935907" [ 850.066689] env[61985]: _type = "Task" [ 850.066689] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.076714] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935907, 'name': Rename_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.319556] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935906, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.425249] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935898, 'name': CloneVM_Task, 'duration_secs': 2.110791} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.429069] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Created linked-clone VM from snapshot [ 850.429069] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb549f8a-a33b-4d4e-9097-42c125448b42 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.436119] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Uploading image f846c12d-4c9e-4d1f-af04-f34584b454cf {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 850.476596] env[61985]: DEBUG oslo_vmware.rw_handles [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 850.476596] env[61985]: value = "vm-211365" [ 850.476596] env[61985]: _type = "VirtualMachine" [ 850.476596] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 850.476993] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-162b8a68-93e9-4048-8f75-6f694da65911 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.484909] env[61985]: DEBUG oslo_vmware.rw_handles [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lease: (returnval){ [ 850.484909] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5266e99a-33d4-aa1e-6216-cb90cb1e68d5" [ 850.484909] env[61985]: _type = "HttpNfcLease" [ 850.484909] env[61985]: } obtained for exporting VM: (result){ [ 850.484909] env[61985]: value = "vm-211365" [ 850.484909] env[61985]: _type = "VirtualMachine" [ 850.484909] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 850.485412] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the lease: (returnval){ [ 850.485412] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5266e99a-33d4-aa1e-6216-cb90cb1e68d5" [ 850.485412] env[61985]: _type = "HttpNfcLease" [ 850.485412] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 850.493185] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 850.493185] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5266e99a-33d4-aa1e-6216-cb90cb1e68d5" [ 850.493185] env[61985]: _type = "HttpNfcLease" [ 850.493185] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 850.547385] env[61985]: DEBUG oslo_vmware.api [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935905, 'name': PowerOnVM_Task, 'duration_secs': 0.533315} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.547897] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 850.549582] env[61985]: INFO nova.compute.manager [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Took 7.08 seconds to spawn the instance on the hypervisor. [ 850.549877] env[61985]: DEBUG nova.compute.manager [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.550879] env[61985]: DEBUG nova.scheduler.client.report [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.555437] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988cf069-39ad-46ab-9044-936d5218c3be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.581678] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935907, 'name': Rename_Task, 'duration_secs': 0.310324} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.581678] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 850.582286] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b05f25b4-6cf9-42bc-86dd-5ffb05fcdac3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.591421] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 850.591421] env[61985]: value = "task-935909" [ 850.591421] env[61985]: _type = "Task" [ 850.591421] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.604311] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935909, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.820484] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935906, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.996358] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 850.996358] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5266e99a-33d4-aa1e-6216-cb90cb1e68d5" [ 850.996358] env[61985]: _type = "HttpNfcLease" [ 850.996358] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 850.997118] env[61985]: DEBUG oslo_vmware.rw_handles [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 850.997118] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5266e99a-33d4-aa1e-6216-cb90cb1e68d5" [ 850.997118] env[61985]: _type = "HttpNfcLease" [ 850.997118] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 850.998206] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85837cc2-7d34-4a55-b56a-b57e8c7b6a0b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.008235] env[61985]: DEBUG oslo_vmware.rw_handles [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523a268d-2c6e-d618-6d57-f05711daf0ab/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 851.008759] env[61985]: DEBUG oslo_vmware.rw_handles [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523a268d-2c6e-d618-6d57-f05711daf0ab/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 851.083379] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.260s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.092442] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.931s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.095103] env[61985]: INFO nova.compute.claims [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.105905] env[61985]: INFO nova.compute.manager [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Took 39.88 seconds to build instance. [ 851.115077] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935909, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.132088] env[61985]: INFO nova.scheduler.client.report [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted allocations for instance f8bc8b71-0317-479d-b2f9-9471757f0774 [ 851.150550] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-48e337df-bf51-4bf4-9595-87727eed20c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.318376] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935906, 'name': CloneVM_Task, 'duration_secs': 1.309461} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.318946] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Created linked-clone VM from snapshot [ 851.320087] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3de3500-5653-4d38-b45b-83f2e87c13ef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.328174] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Uploading image f3f868ff-4e58-483f-92fc-670d25c6792c {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 851.386748] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 851.386838] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c90b47ca-1323-4fcb-b486-a9c12a42243f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.393994] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 851.393994] env[61985]: value = "task-935910" [ 851.393994] env[61985]: _type = "Task" [ 851.393994] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.404736] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935910, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.611191] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6b125696-fa16-4b5a-8917-fbcad037d1a1 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lock "f8924f18-7232-4be6-84bf-1ba05b855ffe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.748s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.624734] env[61985]: DEBUG oslo_vmware.api [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935909, 'name': PowerOnVM_Task, 'duration_secs': 0.696338} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.626203] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 851.627012] env[61985]: INFO nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Took 10.79 seconds to spawn the instance on the hypervisor. [ 851.627733] env[61985]: DEBUG nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.630599] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cc4481-fe42-498e-8355-a654f12039af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.642515] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a329ef47-569d-4cd6-a534-c51ebce775ef tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "f8bc8b71-0317-479d-b2f9-9471757f0774" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.535s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.904336] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935910, 'name': Destroy_Task, 'duration_secs': 0.484514} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.904847] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Destroyed the VM [ 851.905107] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 851.905475] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-649a9520-4f18-4fea-9388-8c20ee26ea54 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.912806] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 851.912806] env[61985]: value = "task-935911" [ 851.912806] env[61985]: _type = "Task" [ 851.912806] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.922551] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935911, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.115489] env[61985]: DEBUG nova.compute.manager [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 852.169219] env[61985]: INFO nova.compute.manager [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Took 42.26 seconds to build instance. [ 852.322879] env[61985]: DEBUG nova.compute.manager [None req-c6bb3cf2-a368-418c-86d3-03db69734e38 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.332180] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1363b7b6-1a9a-448a-ba11-1ab6c21abe30 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.429742] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935911, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.559630] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquiring lock "f8924f18-7232-4be6-84bf-1ba05b855ffe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.559909] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lock "f8924f18-7232-4be6-84bf-1ba05b855ffe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.560162] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquiring lock "f8924f18-7232-4be6-84bf-1ba05b855ffe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.560367] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lock "f8924f18-7232-4be6-84bf-1ba05b855ffe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.561169] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lock "f8924f18-7232-4be6-84bf-1ba05b855ffe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.562803] env[61985]: INFO nova.compute.manager [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Terminating instance [ 852.566948] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquiring lock "refresh_cache-f8924f18-7232-4be6-84bf-1ba05b855ffe" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.567139] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquired lock "refresh_cache-f8924f18-7232-4be6-84bf-1ba05b855ffe" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.567322] env[61985]: DEBUG nova.network.neutron [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.645863] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1eb6e9-1a86-49c0-8f16-157a9ffba3cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.651387] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.655626] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efadc3e-f4af-4299-8545-346117beb2ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.692159] env[61985]: DEBUG oslo_concurrency.lockutils [None req-03cf774c-cbaa-4061-98a7-3117818457b9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.418s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.694813] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5011944-8f3e-4f4f-b728-72e99b481b6b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.704963] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33628bb-9485-4df4-addf-3c33a6864f6c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.720979] env[61985]: DEBUG nova.compute.provider_tree [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.845955] env[61985]: INFO nova.compute.manager [None req-c6bb3cf2-a368-418c-86d3-03db69734e38 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] instance snapshotting [ 852.847443] env[61985]: DEBUG nova.objects.instance [None req-c6bb3cf2-a368-418c-86d3-03db69734e38 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lazy-loading 'flavor' on Instance uuid f8924f18-7232-4be6-84bf-1ba05b855ffe {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.924399] env[61985]: DEBUG oslo_vmware.api [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935911, 'name': RemoveSnapshot_Task, 'duration_secs': 0.892839} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.924724] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 853.088281] env[61985]: DEBUG nova.network.neutron [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.165353] env[61985]: DEBUG nova.network.neutron [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.200096] env[61985]: DEBUG nova.compute.manager [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.223977] env[61985]: DEBUG nova.scheduler.client.report [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.355080] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1439bcd0-38ac-40f8-bf98-64efa60c17db {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.381125] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0558c0-1d6a-4b81-be0c-ae53ad86bce1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.402501] env[61985]: DEBUG oslo_concurrency.lockutils [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "95e3e6d7-9203-47a0-bad7-050eba09c511" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.402501] env[61985]: DEBUG oslo_concurrency.lockutils [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "95e3e6d7-9203-47a0-bad7-050eba09c511" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.402501] env[61985]: DEBUG oslo_concurrency.lockutils [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "95e3e6d7-9203-47a0-bad7-050eba09c511-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.402501] env[61985]: DEBUG oslo_concurrency.lockutils [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "95e3e6d7-9203-47a0-bad7-050eba09c511-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.402747] env[61985]: DEBUG oslo_concurrency.lockutils [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "95e3e6d7-9203-47a0-bad7-050eba09c511-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.403268] env[61985]: INFO nova.compute.manager [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Terminating instance [ 853.409023] env[61985]: DEBUG nova.compute.manager [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 853.409023] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 853.409023] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ea64fc-c526-41ef-a49e-031d5c460762 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.418651] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 853.419078] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2afce939-1eb9-4886-bd7d-e5b615a52eb2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.426747] env[61985]: DEBUG oslo_vmware.api [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 853.426747] env[61985]: value = "task-935912" [ 853.426747] env[61985]: _type = "Task" [ 853.426747] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.436231] env[61985]: WARNING nova.compute.manager [None req-586ffb46-f004-45ab-ba8b-c84945e5f3a2 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Image not found during snapshot: nova.exception.ImageNotFound: Image f3f868ff-4e58-483f-92fc-670d25c6792c could not be found. [ 853.443095] env[61985]: DEBUG oslo_vmware.api [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.669513] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Releasing lock "refresh_cache-f8924f18-7232-4be6-84bf-1ba05b855ffe" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.670122] env[61985]: DEBUG nova.compute.manager [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 853.670439] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 853.671495] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5d7fb0-8d4a-45d8-95e1-2e7623d74f18 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.685144] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 853.685538] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3964f930-fcc6-43d8-b453-65293ad6420b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.694344] env[61985]: DEBUG oslo_vmware.api [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for the task: (returnval){ [ 853.694344] env[61985]: value = "task-935913" [ 853.694344] env[61985]: _type = "Task" [ 853.694344] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.704706] env[61985]: DEBUG oslo_vmware.api [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935913, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.730962] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.735018] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.642s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.735401] env[61985]: DEBUG nova.compute.manager [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 853.738306] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.363s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.740309] env[61985]: INFO nova.compute.claims [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.820931] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.821298] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.821535] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.821737] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.821980] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.825862] env[61985]: INFO nova.compute.manager [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Terminating instance [ 853.831881] env[61985]: DEBUG nova.compute.manager [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 853.833436] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 853.833436] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15e4f2d-faa7-4ad3-9350-a41e1986e7aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.843773] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 853.844142] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e501c462-9ae4-4c87-b4c9-d016ba33674f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.851087] env[61985]: DEBUG oslo_vmware.api [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 853.851087] env[61985]: value = "task-935914" [ 853.851087] env[61985]: _type = "Task" [ 853.851087] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.861424] env[61985]: DEBUG oslo_vmware.api [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935914, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.894985] env[61985]: DEBUG nova.compute.manager [None req-c6bb3cf2-a368-418c-86d3-03db69734e38 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Instance disappeared during snapshot {{(pid=61985) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 853.945107] env[61985]: DEBUG oslo_vmware.api [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935912, 'name': PowerOffVM_Task, 'duration_secs': 0.233035} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.945429] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 853.945613] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 853.945884] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc63facb-381e-4f41-a4e7-2084733df878 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.204823] env[61985]: DEBUG oslo_vmware.api [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935913, 'name': PowerOffVM_Task, 'duration_secs': 0.145574} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.205223] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 854.205414] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 854.205693] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67c14181-cd7f-4394-85f4-260dfd5ebb8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.231845] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 854.232233] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 854.232233] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Deleting the datastore file [datastore2] f8924f18-7232-4be6-84bf-1ba05b855ffe {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.232383] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b8601159-1f62-4130-8759-4c8f92de5a18 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.239062] env[61985]: DEBUG oslo_vmware.api [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for the task: (returnval){ [ 854.239062] env[61985]: value = "task-935917" [ 854.239062] env[61985]: _type = "Task" [ 854.239062] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.247846] env[61985]: DEBUG nova.compute.utils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.254681] env[61985]: DEBUG nova.compute.manager [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 854.254903] env[61985]: DEBUG nova.network.neutron [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 854.256885] env[61985]: DEBUG oslo_vmware.api [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.260044] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 854.260044] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 854.260202] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleting the datastore file [datastore2] 95e3e6d7-9203-47a0-bad7-050eba09c511 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.260435] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-67a998a1-ee13-40d6-9a97-bf63d20d47f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.267788] env[61985]: DEBUG oslo_vmware.api [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 854.267788] env[61985]: value = "task-935918" [ 854.267788] env[61985]: _type = "Task" [ 854.267788] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.277245] env[61985]: DEBUG oslo_vmware.api [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935918, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.316439] env[61985]: DEBUG nova.policy [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f832e586e8b744f89388329f518ed6df', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83e16978f0db4a3683dcdaf82655fc73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.364241] env[61985]: DEBUG oslo_vmware.api [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935914, 'name': PowerOffVM_Task, 'duration_secs': 0.191366} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.364775] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 854.364905] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 854.365305] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-94ea430c-8c32-4346-bac8-44227a8c7231 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.538359] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 854.538359] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 854.538359] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleting the datastore file [datastore2] e82d26b1-5502-4fd2-89c5-ffb0c1557c79 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.538359] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f299bce-5aaa-44b5-9171-ded2d2f4da6a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.545501] env[61985]: DEBUG oslo_vmware.api [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 854.545501] env[61985]: value = "task-935920" [ 854.545501] env[61985]: _type = "Task" [ 854.545501] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.555293] env[61985]: DEBUG oslo_vmware.api [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935920, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.720971] env[61985]: DEBUG nova.network.neutron [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Successfully created port: c4da38a8-666e-4d80-8841-66ef028ba74a {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.751076] env[61985]: DEBUG nova.compute.manager [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 854.756550] env[61985]: DEBUG oslo_vmware.api [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Task: {'id': task-935917, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161686} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.758342] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.758619] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 854.758856] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 854.759359] env[61985]: INFO nova.compute.manager [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Took 1.09 seconds to destroy the instance on the hypervisor. [ 854.759423] env[61985]: DEBUG oslo.service.loopingcall [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.759654] env[61985]: DEBUG nova.compute.manager [-] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.759976] env[61985]: DEBUG nova.network.neutron [-] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 854.781701] env[61985]: DEBUG oslo_vmware.api [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935918, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.261151} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.782544] env[61985]: DEBUG nova.network.neutron [-] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 854.784360] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.784360] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 854.784360] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 854.784490] env[61985]: INFO nova.compute.manager [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Took 1.38 seconds to destroy the instance on the hypervisor. [ 854.784676] env[61985]: DEBUG oslo.service.loopingcall [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.785825] env[61985]: DEBUG nova.compute.manager [-] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.785825] env[61985]: DEBUG nova.network.neutron [-] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 855.056587] env[61985]: DEBUG oslo_vmware.api [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-935920, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176209} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.059397] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.059607] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 855.059788] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 855.060029] env[61985]: INFO nova.compute.manager [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Took 1.23 seconds to destroy the instance on the hypervisor. [ 855.060302] env[61985]: DEBUG oslo.service.loopingcall [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.060719] env[61985]: DEBUG nova.compute.manager [-] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 855.060825] env[61985]: DEBUG nova.network.neutron [-] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 855.142706] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ab7533-296d-4153-bf5b-46e2d6de2e59 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.153037] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be247c61-1095-4036-afb2-4c499d5e4e57 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.199231] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a912164-56ca-492c-ae1c-607d0c15f8e5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.207631] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7443b1-10c0-4b6c-a469-37bca6b9dad1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.227219] env[61985]: DEBUG nova.compute.provider_tree [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.233854] env[61985]: DEBUG nova.network.neutron [-] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.286808] env[61985]: DEBUG nova.network.neutron [-] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.737137] env[61985]: INFO nova.compute.manager [-] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Took 0.95 seconds to deallocate network for instance. [ 855.737137] env[61985]: DEBUG nova.scheduler.client.report [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.765941] env[61985]: DEBUG nova.compute.manager [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 855.789491] env[61985]: INFO nova.compute.manager [-] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Took 1.03 seconds to deallocate network for instance. [ 855.982031] env[61985]: DEBUG nova.network.neutron [-] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.127415] env[61985]: DEBUG nova.compute.manager [req-a4a6cb00-51af-4656-bed9-8b935c46e18c req-7b784a50-faf0-4012-a614-0f7e9a351295 service nova] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Received event network-vif-deleted-d3b72151-261d-424c-adf2-aa8c1b98ec5e {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 856.189015] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.189356] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.189609] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.189825] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.190091] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.192371] env[61985]: INFO nova.compute.manager [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Terminating instance [ 856.194394] env[61985]: DEBUG nova.compute.manager [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 856.194624] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 856.195682] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7584d881-2485-44a3-b6bb-ffb20c09ff47 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.204254] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 856.204553] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-485b5849-4fba-43da-bb5d-7133a2c501f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.210856] env[61985]: DEBUG oslo_vmware.api [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 856.210856] env[61985]: value = "task-935921" [ 856.210856] env[61985]: _type = "Task" [ 856.210856] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.220506] env[61985]: DEBUG oslo_vmware.api [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935921, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.246864] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.248533] env[61985]: DEBUG oslo_concurrency.lockutils [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.682s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.248871] env[61985]: DEBUG nova.objects.instance [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lazy-loading 'resources' on Instance uuid 8f6cd002-b3c2-4276-b195-15d09a143d31 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.250945] env[61985]: DEBUG oslo_concurrency.lockutils [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.296276] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.314217] env[61985]: DEBUG nova.compute.manager [req-83c05c91-2950-438a-b9bc-10c5c5829c45 req-3f1f4190-ea3e-4b3d-926b-eac0563cd208 service nova] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Received event network-vif-deleted-ae5604d8-2112-4be8-a6d8-d5a94b4eac44 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 856.484913] env[61985]: INFO nova.compute.manager [-] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Took 1.42 seconds to deallocate network for instance. [ 856.721661] env[61985]: DEBUG oslo_vmware.api [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935921, 'name': PowerOffVM_Task, 'duration_secs': 0.19826} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.721958] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 856.722432] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 856.722729] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a3631b0-4f80-454c-bc78-3cce973fa22a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.745442] env[61985]: DEBUG nova.network.neutron [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Successfully updated port: c4da38a8-666e-4d80-8841-66ef028ba74a {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.755180] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquiring lock "1c9079c7-5864-4ecf-923e-ab7c7939009d" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.755453] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "1c9079c7-5864-4ecf-923e-ab7c7939009d" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.793242] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 856.793493] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 856.793691] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Deleting the datastore file [datastore1] 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 856.794309] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2fe78182-cd20-495a-bcb8-e3ccea05ac3d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.800532] env[61985]: DEBUG oslo_vmware.api [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 856.800532] env[61985]: value = "task-935923" [ 856.800532] env[61985]: _type = "Task" [ 856.800532] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.811534] env[61985]: DEBUG oslo_vmware.api [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935923, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.994180] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.200993] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb83cf7-e6cb-431e-b520-f1be55071240 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.211251] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ad7db2-4212-48ab-9e47-b7d41724d575 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.243973] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560212ad-409a-4643-b816-31f252f58811 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.247397] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "refresh_cache-1f284789-1e7b-4e9f-9670-34e8e25cd797" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.247573] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquired lock "refresh_cache-1f284789-1e7b-4e9f-9670-34e8e25cd797" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.247723] env[61985]: DEBUG nova.network.neutron [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 857.254814] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69079975-391c-4d05-82f5-ee5c2973bde4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.259683] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "1c9079c7-5864-4ecf-923e-ab7c7939009d" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.504s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.260233] env[61985]: DEBUG nova.compute.manager [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 857.273462] env[61985]: DEBUG nova.compute.provider_tree [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.313909] env[61985]: DEBUG oslo_vmware.api [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-935923, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283645} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.314182] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 857.314373] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 857.314559] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 857.314734] env[61985]: INFO nova.compute.manager [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Took 1.12 seconds to destroy the instance on the hypervisor. [ 857.314977] env[61985]: DEBUG oslo.service.loopingcall [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.315185] env[61985]: DEBUG nova.compute.manager [-] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 857.315282] env[61985]: DEBUG nova.network.neutron [-] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 857.765174] env[61985]: DEBUG nova.compute.utils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.766612] env[61985]: DEBUG nova.compute.manager [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.766792] env[61985]: DEBUG nova.network.neutron [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 857.776255] env[61985]: DEBUG nova.scheduler.client.report [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.804998] env[61985]: DEBUG nova.network.neutron [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 857.845577] env[61985]: DEBUG nova.policy [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a62185e10f104388aaa06d7542ee0214', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '880ca654592f40a383310668368e23d6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 858.034586] env[61985]: DEBUG nova.network.neutron [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Updating instance_info_cache with network_info: [{"id": "c4da38a8-666e-4d80-8841-66ef028ba74a", "address": "fa:16:3e:c2:0c:5a", "network": {"id": "f5bcb96f-b906-42e5-aedd-c0e965806389", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-490613410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83e16978f0db4a3683dcdaf82655fc73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4da38a8-66", "ovs_interfaceid": "c4da38a8-666e-4d80-8841-66ef028ba74a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.068399] env[61985]: DEBUG nova.network.neutron [-] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.178928] env[61985]: DEBUG nova.network.neutron [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Successfully created port: 9669d21e-aa85-402c-81f8-b83cb1368351 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.270446] env[61985]: DEBUG nova.compute.manager [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 858.281249] env[61985]: DEBUG oslo_concurrency.lockutils [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.033s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.283866] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.642s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.286117] env[61985]: INFO nova.compute.claims [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.311236] env[61985]: INFO nova.scheduler.client.report [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted allocations for instance 8f6cd002-b3c2-4276-b195-15d09a143d31 [ 858.348541] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.349337] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.349575] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.349848] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.350081] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.351108] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.351704] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.351840] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.352557] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.352557] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.352557] env[61985]: DEBUG nova.virt.hardware [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.357035] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c5667e-3056-40cd-a0ec-5a2f7ee0fb9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.371225] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edc3fa9-41ba-4e6d-8e82-6db76b36df62 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.429147] env[61985]: DEBUG oslo_vmware.rw_handles [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5239c6fe-8ff3-369f-c621-8095d63dff1e/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 858.430194] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbd1c26-24c9-4f16-add4-878317a25a05 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.437791] env[61985]: DEBUG oslo_vmware.rw_handles [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5239c6fe-8ff3-369f-c621-8095d63dff1e/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 858.438048] env[61985]: ERROR oslo_vmware.rw_handles [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5239c6fe-8ff3-369f-c621-8095d63dff1e/disk-0.vmdk due to incomplete transfer. [ 858.438339] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5a3fdf67-8529-427a-b6d0-b7e41e8e816a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.441197] env[61985]: DEBUG nova.compute.manager [None req-c6bb3cf2-a368-418c-86d3-03db69734e38 tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Found 0 images (rotation: 2) {{(pid=61985) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4560}} [ 858.450869] env[61985]: DEBUG oslo_vmware.rw_handles [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5239c6fe-8ff3-369f-c621-8095d63dff1e/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 858.451149] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Uploaded image 288950a9-1d99-44ba-aa92-2841a4ca12ea to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 858.455119] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 858.455119] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-af739676-e708-4a4b-a521-8a9fa2c88fd8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.461868] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 858.461868] env[61985]: value = "task-935924" [ 858.461868] env[61985]: _type = "Task" [ 858.461868] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.474530] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935924, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.537797] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Releasing lock "refresh_cache-1f284789-1e7b-4e9f-9670-34e8e25cd797" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.538679] env[61985]: DEBUG nova.compute.manager [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Instance network_info: |[{"id": "c4da38a8-666e-4d80-8841-66ef028ba74a", "address": "fa:16:3e:c2:0c:5a", "network": {"id": "f5bcb96f-b906-42e5-aedd-c0e965806389", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-490613410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83e16978f0db4a3683dcdaf82655fc73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4da38a8-66", "ovs_interfaceid": "c4da38a8-666e-4d80-8841-66ef028ba74a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.539209] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:0c:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd8383707-f093-40a7-a5ba-31b0e07cac45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4da38a8-666e-4d80-8841-66ef028ba74a', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.548525] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Creating folder: Project (83e16978f0db4a3683dcdaf82655fc73). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 858.548878] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c5e68ba-280e-46b7-88ac-69f77d2f32f7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.559967] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Created folder: Project (83e16978f0db4a3683dcdaf82655fc73) in parent group-v211285. [ 858.560226] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Creating folder: Instances. Parent ref: group-v211368. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 858.560504] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd9c87ba-f684-4611-878e-9ac28d39581c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.570765] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Created folder: Instances in parent group-v211368. [ 858.571032] env[61985]: DEBUG oslo.service.loopingcall [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.571287] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 858.571525] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d75364ec-4ed4-49ef-942e-4ec3d0340f6f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.588226] env[61985]: INFO nova.compute.manager [-] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Took 1.27 seconds to deallocate network for instance. [ 858.597035] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.597035] env[61985]: value = "task-935927" [ 858.597035] env[61985]: _type = "Task" [ 858.597035] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.608033] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935927, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.826373] env[61985]: DEBUG oslo_concurrency.lockutils [None req-48a2bb75-503f-4fbb-bdc3-f2b06b27013c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8f6cd002-b3c2-4276-b195-15d09a143d31" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.625s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.971958] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935924, 'name': Destroy_Task, 'duration_secs': 0.36702} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.972256] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Destroyed the VM [ 858.972504] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 858.972773] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fe286375-080d-403c-a938-c579f6658592 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.978811] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 858.978811] env[61985]: value = "task-935928" [ 858.978811] env[61985]: _type = "Task" [ 858.978811] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.989373] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935928, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.097528] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.107140] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935927, 'name': CreateVM_Task, 'duration_secs': 0.423527} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.107333] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 859.108152] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.108341] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.108756] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.109087] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff18613c-3d75-4a96-ab70-fc38a0d731a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.119334] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 859.119334] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526639cc-ee8f-c350-a2f7-81d548359cf2" [ 859.119334] env[61985]: _type = "Task" [ 859.119334] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.127856] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526639cc-ee8f-c350-a2f7-81d548359cf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.156611] env[61985]: DEBUG oslo_vmware.rw_handles [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523a268d-2c6e-d618-6d57-f05711daf0ab/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 859.157577] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8bcd8d-4632-4301-82fc-2a6438a7372b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.164178] env[61985]: DEBUG oslo_vmware.rw_handles [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523a268d-2c6e-d618-6d57-f05711daf0ab/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 859.164739] env[61985]: ERROR oslo_vmware.rw_handles [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523a268d-2c6e-d618-6d57-f05711daf0ab/disk-0.vmdk due to incomplete transfer. [ 859.164739] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-56d666b7-ef8e-4776-bc90-d83af6c5e6e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.172095] env[61985]: DEBUG oslo_vmware.rw_handles [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523a268d-2c6e-d618-6d57-f05711daf0ab/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 859.172309] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Uploaded image f846c12d-4c9e-4d1f-af04-f34584b454cf to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 859.173955] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 859.175288] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-81200d3f-0d50-4901-9b68-d9b8530adf2a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.178606] env[61985]: DEBUG nova.compute.manager [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Received event network-vif-plugged-c4da38a8-666e-4d80-8841-66ef028ba74a {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 859.179093] env[61985]: DEBUG oslo_concurrency.lockutils [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] Acquiring lock "1f284789-1e7b-4e9f-9670-34e8e25cd797-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.179329] env[61985]: DEBUG oslo_concurrency.lockutils [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.179506] env[61985]: DEBUG oslo_concurrency.lockutils [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.179681] env[61985]: DEBUG nova.compute.manager [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] No waiting events found dispatching network-vif-plugged-c4da38a8-666e-4d80-8841-66ef028ba74a {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 859.179864] env[61985]: WARNING nova.compute.manager [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Received unexpected event network-vif-plugged-c4da38a8-666e-4d80-8841-66ef028ba74a for instance with vm_state building and task_state spawning. [ 859.180047] env[61985]: DEBUG nova.compute.manager [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Received event network-changed-c4da38a8-666e-4d80-8841-66ef028ba74a {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 859.180215] env[61985]: DEBUG nova.compute.manager [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Refreshing instance network info cache due to event network-changed-c4da38a8-666e-4d80-8841-66ef028ba74a. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 859.180407] env[61985]: DEBUG oslo_concurrency.lockutils [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] Acquiring lock "refresh_cache-1f284789-1e7b-4e9f-9670-34e8e25cd797" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.180550] env[61985]: DEBUG oslo_concurrency.lockutils [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] Acquired lock "refresh_cache-1f284789-1e7b-4e9f-9670-34e8e25cd797" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.180715] env[61985]: DEBUG nova.network.neutron [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Refreshing network info cache for port c4da38a8-666e-4d80-8841-66ef028ba74a {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 859.187144] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 859.187144] env[61985]: value = "task-935929" [ 859.187144] env[61985]: _type = "Task" [ 859.187144] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.196605] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935929, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.281473] env[61985]: DEBUG nova.compute.manager [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 859.309466] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.309613] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.311037] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.311037] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.311037] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.311037] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.311037] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.311240] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.311240] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.311240] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.311408] env[61985]: DEBUG nova.virt.hardware [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.312360] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b20d5fa-3aac-4142-833a-53a6712e6a4d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.326093] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79392027-307e-41b0-86a9-465da7910866 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.481011] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "073a7668-39e6-480d-9350-835a0282b456" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.481096] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "073a7668-39e6-480d-9350-835a0282b456" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.493511] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935928, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.633362] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526639cc-ee8f-c350-a2f7-81d548359cf2, 'name': SearchDatastore_Task, 'duration_secs': 0.009228} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.633914] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.634318] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.634564] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.634786] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.635142] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.637706] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c7061bd6-6fbb-48c6-9d89-88a17857280d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.646091] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.646282] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 859.646996] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08b42b2d-815f-4523-89e3-867158eeb0bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.654346] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 859.654346] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521e37ab-1c17-ab3b-b9ba-b53c764c3178" [ 859.654346] env[61985]: _type = "Task" [ 859.654346] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.662425] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521e37ab-1c17-ab3b-b9ba-b53c764c3178, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.674650] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92eaf95b-c02b-4e74-9bf0-9478708b26ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.682602] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a55abe6-69c5-4614-b997-c0b64412459a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.722081] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65851f8-bceb-4cc8-a4d2-04e4d1fdcd2a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.728188] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935929, 'name': Destroy_Task, 'duration_secs': 0.350371} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.728825] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Destroyed the VM [ 859.729135] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 859.729385] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5dc7c964-e8f8-403f-8681-3c0c33b18b9d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.736693] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685702bd-266f-4d99-ad59-5709f760b01e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.742734] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 859.742734] env[61985]: value = "task-935930" [ 859.742734] env[61985]: _type = "Task" [ 859.742734] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.757207] env[61985]: DEBUG nova.compute.provider_tree [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.768288] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935930, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.990625] env[61985]: DEBUG oslo_vmware.api [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935928, 'name': RemoveSnapshot_Task, 'duration_secs': 0.568415} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.990932] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 859.991792] env[61985]: INFO nova.compute.manager [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Took 17.80 seconds to snapshot the instance on the hypervisor. [ 860.131689] env[61985]: DEBUG nova.network.neutron [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Updated VIF entry in instance network info cache for port c4da38a8-666e-4d80-8841-66ef028ba74a. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 860.132297] env[61985]: DEBUG nova.network.neutron [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Updating instance_info_cache with network_info: [{"id": "c4da38a8-666e-4d80-8841-66ef028ba74a", "address": "fa:16:3e:c2:0c:5a", "network": {"id": "f5bcb96f-b906-42e5-aedd-c0e965806389", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-490613410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83e16978f0db4a3683dcdaf82655fc73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4da38a8-66", "ovs_interfaceid": "c4da38a8-666e-4d80-8841-66ef028ba74a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.164438] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521e37ab-1c17-ab3b-b9ba-b53c764c3178, 'name': SearchDatastore_Task, 'duration_secs': 0.009342} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.165236] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a101d43-66b0-45ee-aae4-d70cfef0a748 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.170362] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 860.170362] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52663055-4083-9783-ad4c-3afbf2f122a4" [ 860.170362] env[61985]: _type = "Task" [ 860.170362] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.176225] env[61985]: DEBUG nova.compute.manager [req-d4eb1848-4ad6-4835-b819-5599c7319021 req-ffbf4b8b-d12c-40d1-a895-576d431c1577 service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Received event network-vif-plugged-9669d21e-aa85-402c-81f8-b83cb1368351 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 860.176809] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4eb1848-4ad6-4835-b819-5599c7319021 req-ffbf4b8b-d12c-40d1-a895-576d431c1577 service nova] Acquiring lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.176874] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4eb1848-4ad6-4835-b819-5599c7319021 req-ffbf4b8b-d12c-40d1-a895-576d431c1577 service nova] Lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.177336] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4eb1848-4ad6-4835-b819-5599c7319021 req-ffbf4b8b-d12c-40d1-a895-576d431c1577 service nova] Lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.177634] env[61985]: DEBUG nova.compute.manager [req-d4eb1848-4ad6-4835-b819-5599c7319021 req-ffbf4b8b-d12c-40d1-a895-576d431c1577 service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] No waiting events found dispatching network-vif-plugged-9669d21e-aa85-402c-81f8-b83cb1368351 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.177885] env[61985]: WARNING nova.compute.manager [req-d4eb1848-4ad6-4835-b819-5599c7319021 req-ffbf4b8b-d12c-40d1-a895-576d431c1577 service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Received unexpected event network-vif-plugged-9669d21e-aa85-402c-81f8-b83cb1368351 for instance with vm_state building and task_state spawning. [ 860.182157] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52663055-4083-9783-ad4c-3afbf2f122a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.253337] env[61985]: DEBUG oslo_vmware.api [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935930, 'name': RemoveSnapshot_Task, 'duration_secs': 0.450337} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.253615] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 860.253844] env[61985]: INFO nova.compute.manager [None req-251f4a7d-61c8-463a-970a-5e8eb914ba4b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Took 14.98 seconds to snapshot the instance on the hypervisor. [ 860.263272] env[61985]: DEBUG nova.scheduler.client.report [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.364192] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.364448] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.538573] env[61985]: DEBUG nova.compute.manager [None req-6ae7095e-23e2-406d-9974-6855496ddf3c tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Found 2 images (rotation: 2) {{(pid=61985) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4560}} [ 860.604285] env[61985]: DEBUG nova.network.neutron [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Successfully updated port: 9669d21e-aa85-402c-81f8-b83cb1368351 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.635409] env[61985]: DEBUG oslo_concurrency.lockutils [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] Releasing lock "refresh_cache-1f284789-1e7b-4e9f-9670-34e8e25cd797" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.635712] env[61985]: DEBUG nova.compute.manager [req-914282ad-1ac1-4bf8-aef2-a7c1046c20fe req-f4cec159-4bdc-4e1c-a4eb-59ad47d19de8 service nova] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Received event network-vif-deleted-5b8f2f2c-0511-4436-982f-3f32c10f5759 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 860.680962] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52663055-4083-9783-ad4c-3afbf2f122a4, 'name': SearchDatastore_Task, 'duration_secs': 0.008663} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.681267] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.681535] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 1f284789-1e7b-4e9f-9670-34e8e25cd797/1f284789-1e7b-4e9f-9670-34e8e25cd797.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 860.681803] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4364e211-9dc1-49d6-953b-0982cf0a3c4e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.689120] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 860.689120] env[61985]: value = "task-935931" [ 860.689120] env[61985]: _type = "Task" [ 860.689120] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.697231] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935931, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.713172] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquiring lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.713420] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.713649] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquiring lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.713883] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.714077] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.716530] env[61985]: INFO nova.compute.manager [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Terminating instance [ 860.718597] env[61985]: DEBUG nova.compute.manager [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 860.718814] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 860.719614] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe629225-7f7e-489e-b016-8b7bbd868d56 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.726975] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 860.727120] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b3d28f0-63e9-4365-9265-cb45b8648da1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.733163] env[61985]: DEBUG oslo_vmware.api [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for the task: (returnval){ [ 860.733163] env[61985]: value = "task-935932" [ 860.733163] env[61985]: _type = "Task" [ 860.733163] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.741059] env[61985]: DEBUG oslo_vmware.api [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935932, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.769064] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.769064] env[61985]: DEBUG nova.compute.manager [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 860.771054] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.261s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.775400] env[61985]: DEBUG nova.objects.instance [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lazy-loading 'resources' on Instance uuid e6a57868-ab6e-45d2-9fa0-94efbb79185b {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.943377] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "c4e95f51-9acc-46e8-9921-e0a85fb38598" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.943377] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "c4e95f51-9acc-46e8-9921-e0a85fb38598" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.943377] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "c4e95f51-9acc-46e8-9921-e0a85fb38598-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.943377] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "c4e95f51-9acc-46e8-9921-e0a85fb38598-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.943643] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "c4e95f51-9acc-46e8-9921-e0a85fb38598-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.945824] env[61985]: INFO nova.compute.manager [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Terminating instance [ 860.947846] env[61985]: DEBUG nova.compute.manager [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 860.948051] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 860.948898] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2894ab01-889e-4fe8-8707-f6adb723c219 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.956862] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 860.957505] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c50febb0-ef98-4538-ab31-c44dbe5acc7a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.022400] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 861.022642] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 861.022853] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleting the datastore file [datastore1] c4e95f51-9acc-46e8-9921-e0a85fb38598 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 861.023145] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e802c24b-fd30-48e8-a9f9-94d136aab429 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.034812] env[61985]: DEBUG oslo_vmware.api [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 861.034812] env[61985]: value = "task-935934" [ 861.034812] env[61985]: _type = "Task" [ 861.034812] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.043730] env[61985]: DEBUG oslo_vmware.api [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935934, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.109737] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquiring lock "refresh_cache-17bdf8ec-13d2-459c-bc8e-db6a274fc27e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.109737] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquired lock "refresh_cache-17bdf8ec-13d2-459c-bc8e-db6a274fc27e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.109991] env[61985]: DEBUG nova.network.neutron [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 861.199974] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935931, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.244494] env[61985]: DEBUG oslo_vmware.api [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935932, 'name': PowerOffVM_Task, 'duration_secs': 0.23739} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.244700] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 861.244872] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 861.245149] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55c237ce-3f13-4785-86c4-8c0e79b3189b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.274561] env[61985]: DEBUG nova.compute.utils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 861.276160] env[61985]: DEBUG nova.compute.manager [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 861.276333] env[61985]: DEBUG nova.network.neutron [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 861.358732] env[61985]: DEBUG nova.policy [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e23fdb33638143409613d9d4f26f199c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe531ec9f60548adb457351e19f8d51b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 861.412926] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 861.413213] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 861.413364] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Deleting the datastore file [datastore2] 4a94a6f9-1c86-4628-aa63-341f2c114e2a {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 861.413656] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0aa6b43a-737c-450f-b56b-6f49802f0edd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.421742] env[61985]: DEBUG oslo_vmware.api [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for the task: (returnval){ [ 861.421742] env[61985]: value = "task-935936" [ 861.421742] env[61985]: _type = "Task" [ 861.421742] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.429692] env[61985]: DEBUG oslo_vmware.api [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935936, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.437672] env[61985]: DEBUG nova.compute.manager [req-08e3054b-6b47-48a9-8667-aa25d10bc2fd req-21bd8d7e-0c2a-461b-810d-44578897d7aa service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Received event network-changed-9669d21e-aa85-402c-81f8-b83cb1368351 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 861.437901] env[61985]: DEBUG nova.compute.manager [req-08e3054b-6b47-48a9-8667-aa25d10bc2fd req-21bd8d7e-0c2a-461b-810d-44578897d7aa service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Refreshing instance network info cache due to event network-changed-9669d21e-aa85-402c-81f8-b83cb1368351. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 861.438125] env[61985]: DEBUG oslo_concurrency.lockutils [req-08e3054b-6b47-48a9-8667-aa25d10bc2fd req-21bd8d7e-0c2a-461b-810d-44578897d7aa service nova] Acquiring lock "refresh_cache-17bdf8ec-13d2-459c-bc8e-db6a274fc27e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.549356] env[61985]: DEBUG oslo_vmware.api [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-935934, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.28065} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.549827] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.550182] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 861.550534] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 861.550897] env[61985]: INFO nova.compute.manager [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Took 0.60 seconds to destroy the instance on the hypervisor. [ 861.551334] env[61985]: DEBUG oslo.service.loopingcall [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.551711] env[61985]: DEBUG nova.compute.manager [-] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 861.551939] env[61985]: DEBUG nova.network.neutron [-] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 861.588291] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "986c576e-be02-48ac-b24c-72edccab25be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.588740] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "986c576e-be02-48ac-b24c-72edccab25be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.657073] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d4430a-ded6-490f-981b-df5737a5729e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.659721] env[61985]: DEBUG nova.compute.manager [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.661081] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2ac781-b706-4ecb-8703-784288b54635 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.666698] env[61985]: DEBUG nova.network.neutron [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 861.672818] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6cd92f8-f5dd-4bd8-b5d7-9775e12313a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.720991] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3c0b6c-438e-45a4-a2ac-8028ba75cd0d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.729691] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935931, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548327} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.732881] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 1f284789-1e7b-4e9f-9670-34e8e25cd797/1f284789-1e7b-4e9f-9670-34e8e25cd797.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 861.733485] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.734060] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f00d8f54-a58b-4214-a4fa-34a842aaf7f0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.737157] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364d31c8-1a20-47ba-abdc-062225906d01 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.753035] env[61985]: DEBUG nova.compute.provider_tree [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.759125] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 861.759125] env[61985]: value = "task-935937" [ 861.759125] env[61985]: _type = "Task" [ 861.759125] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.768657] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935937, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.781205] env[61985]: DEBUG nova.compute.manager [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 861.868727] env[61985]: DEBUG nova.network.neutron [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Updating instance_info_cache with network_info: [{"id": "9669d21e-aa85-402c-81f8-b83cb1368351", "address": "fa:16:3e:2e:f3:d2", "network": {"id": "0456839f-0a3b-4cdd-8c84-971cb1927902", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-288158515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "880ca654592f40a383310668368e23d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9669d21e-aa", "ovs_interfaceid": "9669d21e-aa85-402c-81f8-b83cb1368351", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.931902] env[61985]: DEBUG oslo_vmware.api [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Task: {'id': task-935936, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145433} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.932187] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.932383] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 861.932567] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 861.932749] env[61985]: INFO nova.compute.manager [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Took 1.21 seconds to destroy the instance on the hypervisor. [ 861.933058] env[61985]: DEBUG oslo.service.loopingcall [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.933269] env[61985]: DEBUG nova.compute.manager [-] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 861.933361] env[61985]: DEBUG nova.network.neutron [-] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 862.040413] env[61985]: DEBUG nova.network.neutron [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Successfully created port: 876d546b-a049-4c69-817f-aa6e7d7a2fc2 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.183888] env[61985]: INFO nova.compute.manager [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] instance snapshotting [ 862.184576] env[61985]: DEBUG nova.objects.instance [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'flavor' on Instance uuid 2322ab72-9841-41fb-9d60-2812baabe108 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.261510] env[61985]: DEBUG nova.scheduler.client.report [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.276897] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935937, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072964} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.277188] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.278017] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da3dd19-5b9a-42b7-a39f-639ae9cc3751 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.304850] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 1f284789-1e7b-4e9f-9670-34e8e25cd797/1f284789-1e7b-4e9f-9670-34e8e25cd797.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.304850] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b5218c0-5b90-4489-befa-4d2ff68d542b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.324533] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 862.324533] env[61985]: value = "task-935938" [ 862.324533] env[61985]: _type = "Task" [ 862.324533] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.332791] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935938, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.370901] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Releasing lock "refresh_cache-17bdf8ec-13d2-459c-bc8e-db6a274fc27e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.371253] env[61985]: DEBUG nova.compute.manager [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Instance network_info: |[{"id": "9669d21e-aa85-402c-81f8-b83cb1368351", "address": "fa:16:3e:2e:f3:d2", "network": {"id": "0456839f-0a3b-4cdd-8c84-971cb1927902", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-288158515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "880ca654592f40a383310668368e23d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9669d21e-aa", "ovs_interfaceid": "9669d21e-aa85-402c-81f8-b83cb1368351", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.371676] env[61985]: DEBUG oslo_concurrency.lockutils [req-08e3054b-6b47-48a9-8667-aa25d10bc2fd req-21bd8d7e-0c2a-461b-810d-44578897d7aa service nova] Acquired lock "refresh_cache-17bdf8ec-13d2-459c-bc8e-db6a274fc27e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.371782] env[61985]: DEBUG nova.network.neutron [req-08e3054b-6b47-48a9-8667-aa25d10bc2fd req-21bd8d7e-0c2a-461b-810d-44578897d7aa service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Refreshing network info cache for port 9669d21e-aa85-402c-81f8-b83cb1368351 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 862.376022] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:f3:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '557aba95-8968-407a-bac2-2fae66f7c8e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9669d21e-aa85-402c-81f8-b83cb1368351', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.380883] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Creating folder: Project (880ca654592f40a383310668368e23d6). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 862.381939] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5dbfc67a-959b-4dbb-9510-73e99a63ed13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.393576] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Created folder: Project (880ca654592f40a383310668368e23d6) in parent group-v211285. [ 862.393895] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Creating folder: Instances. Parent ref: group-v211371. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 862.394150] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-665bbf51-81d1-4781-a45d-d0e47cba551b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.403322] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Created folder: Instances in parent group-v211371. [ 862.403567] env[61985]: DEBUG oslo.service.loopingcall [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.403759] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 862.403966] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ece8119-13a1-4597-99d0-8cbc3b790730 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.428557] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.428557] env[61985]: value = "task-935941" [ 862.428557] env[61985]: _type = "Task" [ 862.428557] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.438507] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935941, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.691277] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6a4003-48e2-4ae4-936f-5607836f3224 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.715189] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aadbfb7-b9c6-437e-b597-b1df20dcfb52 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.770679] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.999s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.774305] env[61985]: DEBUG oslo_concurrency.lockutils [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 29.911s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.799171] env[61985]: DEBUG nova.compute.manager [req-42bfb86f-8c24-4b14-9de9-567fe02e48ae req-1cf76839-24c1-43ff-880a-26a2a889e806 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Received event network-vif-deleted-f734b929-a7bc-4469-b204-3b5afd314844 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 862.799443] env[61985]: INFO nova.compute.manager [req-42bfb86f-8c24-4b14-9de9-567fe02e48ae req-1cf76839-24c1-43ff-880a-26a2a889e806 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Neutron deleted interface f734b929-a7bc-4469-b204-3b5afd314844; detaching it from the instance and deleting it from the info cache [ 862.799802] env[61985]: DEBUG nova.network.neutron [req-42bfb86f-8c24-4b14-9de9-567fe02e48ae req-1cf76839-24c1-43ff-880a-26a2a889e806 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.806948] env[61985]: DEBUG nova.compute.manager [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 862.812777] env[61985]: INFO nova.scheduler.client.report [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Deleted allocations for instance e6a57868-ab6e-45d2-9fa0-94efbb79185b [ 862.836038] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935938, 'name': ReconfigVM_Task, 'duration_secs': 0.328733} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.838329] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 1f284789-1e7b-4e9f-9670-34e8e25cd797/1f284789-1e7b-4e9f-9670-34e8e25cd797.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.840178] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ceff86ad-8ba2-4e21-98b5-292e9c98c61a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.841157] env[61985]: DEBUG nova.network.neutron [-] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.846875] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.847894] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.847894] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.847894] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.847894] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.847894] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.848298] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.848298] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.848530] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.848703] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.849045] env[61985]: DEBUG nova.virt.hardware [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.850100] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1eac0e-adc1-48ca-b28f-2111fd0a157e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.855628] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 862.855628] env[61985]: value = "task-935942" [ 862.855628] env[61985]: _type = "Task" [ 862.855628] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.862437] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83393b5d-da7b-420e-95bb-48dbdb8768fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.869949] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935942, 'name': Rename_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.924257] env[61985]: DEBUG nova.network.neutron [-] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.938769] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935941, 'name': CreateVM_Task, 'duration_secs': 0.433611} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.938926] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 862.939636] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.939819] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.940208] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.940466] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae77ad49-aab3-4370-a466-a57c7ba9d5c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.946630] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for the task: (returnval){ [ 862.946630] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52666606-a68a-30fe-5b99-961e9d8f22de" [ 862.946630] env[61985]: _type = "Task" [ 862.946630] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.954993] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52666606-a68a-30fe-5b99-961e9d8f22de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.115863] env[61985]: DEBUG nova.network.neutron [req-08e3054b-6b47-48a9-8667-aa25d10bc2fd req-21bd8d7e-0c2a-461b-810d-44578897d7aa service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Updated VIF entry in instance network info cache for port 9669d21e-aa85-402c-81f8-b83cb1368351. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 863.116262] env[61985]: DEBUG nova.network.neutron [req-08e3054b-6b47-48a9-8667-aa25d10bc2fd req-21bd8d7e-0c2a-461b-810d-44578897d7aa service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Updating instance_info_cache with network_info: [{"id": "9669d21e-aa85-402c-81f8-b83cb1368351", "address": "fa:16:3e:2e:f3:d2", "network": {"id": "0456839f-0a3b-4cdd-8c84-971cb1927902", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-288158515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "880ca654592f40a383310668368e23d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "557aba95-8968-407a-bac2-2fae66f7c8e5", "external-id": "nsx-vlan-transportzone-45", "segmentation_id": 45, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9669d21e-aa", "ovs_interfaceid": "9669d21e-aa85-402c-81f8-b83cb1368351", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.230547] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 863.230882] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9d5f2db9-fc0e-45c8-8a22-8f9aae9839b6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.238787] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 863.238787] env[61985]: value = "task-935943" [ 863.238787] env[61985]: _type = "Task" [ 863.238787] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.246604] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935943, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.304046] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-686f23af-d8a9-4aeb-a571-d86f3c33b32b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.312617] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d403acba-a8c3-4f70-b03a-7644ccf71894 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.329833] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ac1eac3-5a75-4329-88e0-7d69469c541d tempest-FloatingIPsAssociationNegativeTestJSON-1368232930 tempest-FloatingIPsAssociationNegativeTestJSON-1368232930-project-member] Lock "e6a57868-ab6e-45d2-9fa0-94efbb79185b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.898s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.345799] env[61985]: INFO nova.compute.manager [-] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Took 1.79 seconds to deallocate network for instance. [ 863.346207] env[61985]: DEBUG nova.compute.manager [req-42bfb86f-8c24-4b14-9de9-567fe02e48ae req-1cf76839-24c1-43ff-880a-26a2a889e806 service nova] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Detach interface failed, port_id=f734b929-a7bc-4469-b204-3b5afd314844, reason: Instance 4a94a6f9-1c86-4628-aa63-341f2c114e2a could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 863.372758] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935942, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.427333] env[61985]: INFO nova.compute.manager [-] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Took 1.49 seconds to deallocate network for instance. [ 863.461563] env[61985]: DEBUG nova.compute.manager [req-24eff4e4-aaba-4497-aad3-dfeffab65235 req-2dc128b4-543c-4d01-8375-cf3656a71350 service nova] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Received event network-vif-deleted-876723dc-2a70-4f56-bff6-e8fd73c0a346 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 863.462280] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52666606-a68a-30fe-5b99-961e9d8f22de, 'name': SearchDatastore_Task, 'duration_secs': 0.01042} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.465123] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.465375] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.465630] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.465738] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.465925] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.467175] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f3acfb0-b772-41c6-a9e2-600762fdd846 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.478931] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.479136] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 863.479991] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f70d74b8-4f6b-4bfc-bfe0-ac930f1f713f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.485239] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for the task: (returnval){ [ 863.485239] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f6d403-ffaa-0ad1-7d4c-45d45e643abf" [ 863.485239] env[61985]: _type = "Task" [ 863.485239] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.495958] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f6d403-ffaa-0ad1-7d4c-45d45e643abf, 'name': SearchDatastore_Task} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.496914] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1054d8e-4d76-44fa-987f-222b8823a2da {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.503259] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for the task: (returnval){ [ 863.503259] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fe2cbd-744b-bcd5-d2df-48867a8c9f7b" [ 863.503259] env[61985]: _type = "Task" [ 863.503259] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.511100] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fe2cbd-744b-bcd5-d2df-48867a8c9f7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.619919] env[61985]: DEBUG oslo_concurrency.lockutils [req-08e3054b-6b47-48a9-8667-aa25d10bc2fd req-21bd8d7e-0c2a-461b-810d-44578897d7aa service nova] Releasing lock "refresh_cache-17bdf8ec-13d2-459c-bc8e-db6a274fc27e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.673136] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2acd779c-c617-4027-8d62-2bbe75c717d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.681699] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a6ecd6-1c4c-4d5f-9f9d-ef9361f6cff2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.710410] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f91e4a7-5c0e-46c6-af06-cea0cb02c815 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.717709] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6888362a-46fd-46b1-b414-6add586dc6c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.732470] env[61985]: DEBUG nova.compute.provider_tree [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.747201] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935943, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.854839] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.866260] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935942, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.932945] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.013116] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fe2cbd-744b-bcd5-d2df-48867a8c9f7b, 'name': SearchDatastore_Task, 'duration_secs': 0.009286} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.013580] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.013835] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 17bdf8ec-13d2-459c-bc8e-db6a274fc27e/17bdf8ec-13d2-459c-bc8e-db6a274fc27e.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 864.014211] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d29d9fd-534e-415a-8905-cd522ab2d311 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.021739] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for the task: (returnval){ [ 864.021739] env[61985]: value = "task-935944" [ 864.021739] env[61985]: _type = "Task" [ 864.021739] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.031472] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.235703] env[61985]: DEBUG nova.scheduler.client.report [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.252785] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935943, 'name': CreateSnapshot_Task, 'duration_secs': 0.57364} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.252785] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 864.253134] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f5a7c8-864f-4ba8-afd1-6c790a68f4cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.308193] env[61985]: DEBUG nova.network.neutron [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Successfully updated port: 876d546b-a049-4c69-817f-aa6e7d7a2fc2 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 864.369062] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935942, 'name': Rename_Task, 'duration_secs': 1.171552} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.369446] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 864.369757] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77ec28ca-d939-4012-8ed4-bd45b2bbc740 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.378600] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 864.378600] env[61985]: value = "task-935945" [ 864.378600] env[61985]: _type = "Task" [ 864.378600] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.387742] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935945, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.532013] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935944, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.773715] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 864.774345] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bb2c2b92-4718-4af3-a187-8073bffa8690 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.783449] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 864.783449] env[61985]: value = "task-935946" [ 864.783449] env[61985]: _type = "Task" [ 864.783449] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.792653] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935946, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.810791] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "refresh_cache-59e742d9-c51a-47dd-9ef2-2f070eae94c4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.810950] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "refresh_cache-59e742d9-c51a-47dd-9ef2-2f070eae94c4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.811148] env[61985]: DEBUG nova.network.neutron [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.894928] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935945, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.037824] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.520059} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.038241] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 17bdf8ec-13d2-459c-bc8e-db6a274fc27e/17bdf8ec-13d2-459c-bc8e-db6a274fc27e.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 865.038590] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.038970] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e4da3499-8b20-4798-bb44-bf49ca52c2c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.048346] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for the task: (returnval){ [ 865.048346] env[61985]: value = "task-935947" [ 865.048346] env[61985]: _type = "Task" [ 865.048346] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.061620] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935947, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.251125] env[61985]: DEBUG oslo_concurrency.lockutils [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.477s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.254543] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.933s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.256224] env[61985]: INFO nova.compute.claims [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.296089] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935946, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.362221] env[61985]: DEBUG nova.network.neutron [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 865.390776] env[61985]: DEBUG oslo_vmware.api [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-935945, 'name': PowerOnVM_Task, 'duration_secs': 0.565327} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.391800] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 865.391800] env[61985]: INFO nova.compute.manager [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Took 9.63 seconds to spawn the instance on the hypervisor. [ 865.391800] env[61985]: DEBUG nova.compute.manager [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.392302] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67aa3868-cbcc-4c19-99fd-3de2483c3b89 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.559574] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935947, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.270564} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.559962] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.561092] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311af53e-7366-46ae-a927-da333ed338c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.586205] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 17bdf8ec-13d2-459c-bc8e-db6a274fc27e/17bdf8ec-13d2-459c-bc8e-db6a274fc27e.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.586498] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d18b005-4a5c-4180-99e9-360752b0a313 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.611465] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for the task: (returnval){ [ 865.611465] env[61985]: value = "task-935948" [ 865.611465] env[61985]: _type = "Task" [ 865.611465] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.622128] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935948, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.634307] env[61985]: DEBUG nova.compute.manager [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Received event network-vif-plugged-876d546b-a049-4c69-817f-aa6e7d7a2fc2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 865.634307] env[61985]: DEBUG oslo_concurrency.lockutils [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] Acquiring lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.634399] env[61985]: DEBUG oslo_concurrency.lockutils [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] Lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.634552] env[61985]: DEBUG oslo_concurrency.lockutils [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] Lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.634719] env[61985]: DEBUG nova.compute.manager [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] No waiting events found dispatching network-vif-plugged-876d546b-a049-4c69-817f-aa6e7d7a2fc2 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.634894] env[61985]: WARNING nova.compute.manager [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Received unexpected event network-vif-plugged-876d546b-a049-4c69-817f-aa6e7d7a2fc2 for instance with vm_state building and task_state spawning. [ 865.635223] env[61985]: DEBUG nova.compute.manager [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Received event network-changed-876d546b-a049-4c69-817f-aa6e7d7a2fc2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 865.636136] env[61985]: DEBUG nova.compute.manager [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Refreshing instance network info cache due to event network-changed-876d546b-a049-4c69-817f-aa6e7d7a2fc2. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 865.636136] env[61985]: DEBUG oslo_concurrency.lockutils [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] Acquiring lock "refresh_cache-59e742d9-c51a-47dd-9ef2-2f070eae94c4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.675327] env[61985]: DEBUG nova.network.neutron [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Updating instance_info_cache with network_info: [{"id": "876d546b-a049-4c69-817f-aa6e7d7a2fc2", "address": "fa:16:3e:68:e0:60", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap876d546b-a0", "ovs_interfaceid": "876d546b-a049-4c69-817f-aa6e7d7a2fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.798058] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935946, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.821750] env[61985]: INFO nova.scheduler.client.report [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Deleted allocation for migration feff90ec-33ae-4a19-ab75-42f6fbb33485 [ 865.912751] env[61985]: INFO nova.compute.manager [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Took 49.77 seconds to build instance. [ 866.122587] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935948, 'name': ReconfigVM_Task, 'duration_secs': 0.322435} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.122909] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 17bdf8ec-13d2-459c-bc8e-db6a274fc27e/17bdf8ec-13d2-459c-bc8e-db6a274fc27e.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.123586] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-baa0927d-0dc0-4428-8968-3a544ee4a404 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.130580] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for the task: (returnval){ [ 866.130580] env[61985]: value = "task-935949" [ 866.130580] env[61985]: _type = "Task" [ 866.130580] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.141662] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935949, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.180016] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "refresh_cache-59e742d9-c51a-47dd-9ef2-2f070eae94c4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.180424] env[61985]: DEBUG nova.compute.manager [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Instance network_info: |[{"id": "876d546b-a049-4c69-817f-aa6e7d7a2fc2", "address": "fa:16:3e:68:e0:60", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap876d546b-a0", "ovs_interfaceid": "876d546b-a049-4c69-817f-aa6e7d7a2fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 866.182154] env[61985]: DEBUG oslo_concurrency.lockutils [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] Acquired lock "refresh_cache-59e742d9-c51a-47dd-9ef2-2f070eae94c4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.182154] env[61985]: DEBUG nova.network.neutron [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Refreshing network info cache for port 876d546b-a049-4c69-817f-aa6e7d7a2fc2 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 866.183245] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:e0:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55520f67-d092-4eb7-940f-d7cceaa1ca1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '876d546b-a049-4c69-817f-aa6e7d7a2fc2', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.192393] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Creating folder: Project (fe531ec9f60548adb457351e19f8d51b). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 866.193069] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7ceb733-87a7-4cb3-91a6-0001bd4f91fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.209615] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Created folder: Project (fe531ec9f60548adb457351e19f8d51b) in parent group-v211285. [ 866.209615] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Creating folder: Instances. Parent ref: group-v211376. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 866.209615] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f68c943-43db-4fb1-8225-7efb7016477e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.220656] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Created folder: Instances in parent group-v211376. [ 866.223016] env[61985]: DEBUG oslo.service.loopingcall [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.223016] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 866.223016] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e28ee1a-1d26-4150-b1fe-624400388059 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.244893] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.244893] env[61985]: value = "task-935952" [ 866.244893] env[61985]: _type = "Task" [ 866.244893] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.254551] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935952, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.305948] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935946, 'name': CloneVM_Task, 'duration_secs': 1.500115} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.306362] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Created linked-clone VM from snapshot [ 866.307659] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a545c778-067b-4a63-8af9-be55a2dd07eb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.321061] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Uploading image 9872176a-7a1f-42a1-895a-258479200089 {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 866.332982] env[61985]: DEBUG oslo_concurrency.lockutils [None req-721f228f-f24f-4a1d-a9bc-30355f98ef8b tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 37.031s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.363520] env[61985]: DEBUG oslo_vmware.rw_handles [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 866.363520] env[61985]: value = "vm-211375" [ 866.363520] env[61985]: _type = "VirtualMachine" [ 866.363520] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 866.364360] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-13dfdc40-16b2-4653-8dfe-c50a72d885f4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.374516] env[61985]: DEBUG oslo_vmware.rw_handles [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lease: (returnval){ [ 866.374516] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e312b8-4e59-6bf2-3b40-637008a3d477" [ 866.374516] env[61985]: _type = "HttpNfcLease" [ 866.374516] env[61985]: } obtained for exporting VM: (result){ [ 866.374516] env[61985]: value = "vm-211375" [ 866.374516] env[61985]: _type = "VirtualMachine" [ 866.374516] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 866.374836] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the lease: (returnval){ [ 866.374836] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e312b8-4e59-6bf2-3b40-637008a3d477" [ 866.374836] env[61985]: _type = "HttpNfcLease" [ 866.374836] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 866.384733] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 866.384733] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e312b8-4e59-6bf2-3b40-637008a3d477" [ 866.384733] env[61985]: _type = "HttpNfcLease" [ 866.384733] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 866.414364] env[61985]: DEBUG oslo_concurrency.lockutils [None req-857f6b7e-4f76-4bb4-b000-dd854c23a352 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.051s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.651016] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935949, 'name': Rename_Task, 'duration_secs': 0.159051} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.652292] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 866.653665] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4ce2b2e-16d0-45e3-93e0-fd3a0ef7dac9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.666160] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for the task: (returnval){ [ 866.666160] env[61985]: value = "task-935954" [ 866.666160] env[61985]: _type = "Task" [ 866.666160] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.678034] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.702345] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4f37aa-f540-47fc-af83-7576ea6cfcba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.714272] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1c31de-c411-465e-a7b9-6c20bb6f645e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.749857] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf6d8a0-d338-4ea6-b1df-337e2d38a30a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.759170] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935952, 'name': CreateVM_Task, 'duration_secs': 0.376693} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.763523] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 866.763880] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.763948] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.764630] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 866.765698] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-febf9142-5072-4163-9ab9-db57a5c59118 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.769818] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fc7e598-d294-4fb3-94f3-8a42fb5444da {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.784574] env[61985]: DEBUG nova.compute.provider_tree [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.786337] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 866.786337] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520cf6bb-0ab1-cdc6-5962-6d7d25c75042" [ 866.786337] env[61985]: _type = "Task" [ 866.786337] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.797023] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520cf6bb-0ab1-cdc6-5962-6d7d25c75042, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.886703] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 866.886703] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e312b8-4e59-6bf2-3b40-637008a3d477" [ 866.886703] env[61985]: _type = "HttpNfcLease" [ 866.886703] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 866.886899] env[61985]: DEBUG oslo_vmware.rw_handles [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 866.886899] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e312b8-4e59-6bf2-3b40-637008a3d477" [ 866.886899] env[61985]: _type = "HttpNfcLease" [ 866.886899] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 866.887604] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a56be1f-ac32-4c19-9808-9ea68f5c5794 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.895779] env[61985]: DEBUG oslo_vmware.rw_handles [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52071d0e-6b7c-83f0-a4c5-08f063edab51/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 866.895990] env[61985]: DEBUG oslo_vmware.rw_handles [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52071d0e-6b7c-83f0-a4c5-08f063edab51/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 866.962599] env[61985]: DEBUG nova.compute.manager [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.996766] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c51a35ee-c3d8-4447-9bc2-67f3334dedac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.178183] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935954, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.295532] env[61985]: DEBUG nova.scheduler.client.report [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.311572] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520cf6bb-0ab1-cdc6-5962-6d7d25c75042, 'name': SearchDatastore_Task, 'duration_secs': 0.023331} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.312369] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.312369] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.312429] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.312815] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.313535] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.314461] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49188c41-e9b0-486e-9525-19e2cc7ebc81 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.320876] env[61985]: DEBUG nova.network.neutron [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Updated VIF entry in instance network info cache for port 876d546b-a049-4c69-817f-aa6e7d7a2fc2. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 867.321431] env[61985]: DEBUG nova.network.neutron [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Updating instance_info_cache with network_info: [{"id": "876d546b-a049-4c69-817f-aa6e7d7a2fc2", "address": "fa:16:3e:68:e0:60", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap876d546b-a0", "ovs_interfaceid": "876d546b-a049-4c69-817f-aa6e7d7a2fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.334980] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.335440] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 867.336928] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d19c76a2-1c40-4b21-95a7-f6d5c07ebc5f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.344541] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 867.344541] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d02285-68d0-7529-3694-b199196dee26" [ 867.344541] env[61985]: _type = "Task" [ 867.344541] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.355219] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d02285-68d0-7529-3694-b199196dee26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.499205] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.684314] env[61985]: DEBUG oslo_vmware.api [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935954, 'name': PowerOnVM_Task, 'duration_secs': 0.638178} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.684640] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 867.684883] env[61985]: INFO nova.compute.manager [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Took 8.40 seconds to spawn the instance on the hypervisor. [ 867.686018] env[61985]: DEBUG nova.compute.manager [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.686018] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8143a7a1-c537-4f71-9c45-a65c58685952 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.804157] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.804157] env[61985]: DEBUG nova.compute.manager [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 867.807480] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.401s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.812719] env[61985]: INFO nova.compute.claims [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.825853] env[61985]: DEBUG oslo_concurrency.lockutils [req-1a46c2bc-1912-4a54-bb47-8c795a5879a5 req-891ff2b0-2b0b-4dd8-a018-e94106b96b11 service nova] Releasing lock "refresh_cache-59e742d9-c51a-47dd-9ef2-2f070eae94c4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.867532] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d02285-68d0-7529-3694-b199196dee26, 'name': SearchDatastore_Task, 'duration_secs': 0.016617} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.867532] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f26c908c-9933-4462-8b1e-623a4c687e82 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.875336] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 867.875336] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a88414-2ae3-07c2-41ac-3104b0f6c351" [ 867.875336] env[61985]: _type = "Task" [ 867.875336] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.886978] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a88414-2ae3-07c2-41ac-3104b0f6c351, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.964141] env[61985]: DEBUG nova.compute.manager [req-d309964e-efe6-40b1-a0b7-44ac437d340d req-3624fe82-86d8-41af-b2b8-c4769d0b8200 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Received event network-changed-c4da38a8-666e-4d80-8841-66ef028ba74a {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 867.964141] env[61985]: DEBUG nova.compute.manager [req-d309964e-efe6-40b1-a0b7-44ac437d340d req-3624fe82-86d8-41af-b2b8-c4769d0b8200 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Refreshing instance network info cache due to event network-changed-c4da38a8-666e-4d80-8841-66ef028ba74a. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 867.964141] env[61985]: DEBUG oslo_concurrency.lockutils [req-d309964e-efe6-40b1-a0b7-44ac437d340d req-3624fe82-86d8-41af-b2b8-c4769d0b8200 service nova] Acquiring lock "refresh_cache-1f284789-1e7b-4e9f-9670-34e8e25cd797" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.964141] env[61985]: DEBUG oslo_concurrency.lockutils [req-d309964e-efe6-40b1-a0b7-44ac437d340d req-3624fe82-86d8-41af-b2b8-c4769d0b8200 service nova] Acquired lock "refresh_cache-1f284789-1e7b-4e9f-9670-34e8e25cd797" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.964956] env[61985]: DEBUG nova.network.neutron [req-d309964e-efe6-40b1-a0b7-44ac437d340d req-3624fe82-86d8-41af-b2b8-c4769d0b8200 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Refreshing network info cache for port c4da38a8-666e-4d80-8841-66ef028ba74a {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 868.208127] env[61985]: INFO nova.compute.manager [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Took 50.85 seconds to build instance. [ 868.321306] env[61985]: DEBUG nova.compute.utils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.325153] env[61985]: DEBUG nova.compute.manager [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.327691] env[61985]: DEBUG nova.network.neutron [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 868.389222] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a88414-2ae3-07c2-41ac-3104b0f6c351, 'name': SearchDatastore_Task, 'duration_secs': 0.015639} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.389565] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.389830] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 59e742d9-c51a-47dd-9ef2-2f070eae94c4/59e742d9-c51a-47dd-9ef2-2f070eae94c4.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 868.390543] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90ba6510-6cfb-4a11-8ceb-19e32b843aa5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.399427] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 868.399427] env[61985]: value = "task-935955" [ 868.399427] env[61985]: _type = "Task" [ 868.399427] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.406916] env[61985]: DEBUG nova.policy [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4f5b5ce76e8b4fae9c905bc7a39b683e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4923009773c047519b883745c59e3516', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 868.413429] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.623187] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.623598] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.711458] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b547c10c-73c7-408d-89fa-4d19829087bf tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.697s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.826310] env[61985]: DEBUG nova.compute.manager [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 868.831262] env[61985]: DEBUG nova.network.neutron [req-d309964e-efe6-40b1-a0b7-44ac437d340d req-3624fe82-86d8-41af-b2b8-c4769d0b8200 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Updated VIF entry in instance network info cache for port c4da38a8-666e-4d80-8841-66ef028ba74a. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 868.831998] env[61985]: DEBUG nova.network.neutron [req-d309964e-efe6-40b1-a0b7-44ac437d340d req-3624fe82-86d8-41af-b2b8-c4769d0b8200 service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Updating instance_info_cache with network_info: [{"id": "c4da38a8-666e-4d80-8841-66ef028ba74a", "address": "fa:16:3e:c2:0c:5a", "network": {"id": "f5bcb96f-b906-42e5-aedd-c0e965806389", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-490613410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83e16978f0db4a3683dcdaf82655fc73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4da38a8-66", "ovs_interfaceid": "c4da38a8-666e-4d80-8841-66ef028ba74a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.912530] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935955, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.216706] env[61985]: DEBUG nova.compute.manager [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 869.280465] env[61985]: DEBUG nova.network.neutron [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Successfully created port: 4d18bb14-4f82-46b7-b128-9acc3eaae43f {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.341225] env[61985]: DEBUG oslo_concurrency.lockutils [req-d309964e-efe6-40b1-a0b7-44ac437d340d req-3624fe82-86d8-41af-b2b8-c4769d0b8200 service nova] Releasing lock "refresh_cache-1f284789-1e7b-4e9f-9670-34e8e25cd797" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.343648] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3d43d8-aa0a-4695-be9e-80ca13ed42e3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.354622] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506a33d0-1d2a-406e-9f2b-37d9d9e0a1a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.395082] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec3fb3f-5dcc-4dcc-a8d1-9a510dbf7579 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.411136] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a218306-8f1e-4480-9ab4-2b0bc5856807 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.419898] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935955, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.713052} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.420788] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 59e742d9-c51a-47dd-9ef2-2f070eae94c4/59e742d9-c51a-47dd-9ef2-2f070eae94c4.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 869.421140] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 869.421489] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-833d3883-68ca-4f0a-9c0c-c11c744a37bc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.433481] env[61985]: DEBUG nova.compute.provider_tree [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.441385] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 869.441385] env[61985]: value = "task-935956" [ 869.441385] env[61985]: _type = "Task" [ 869.441385] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.455982] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935956, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.630091] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquiring lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.630274] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.630505] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquiring lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.630716] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.630937] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.636098] env[61985]: INFO nova.compute.manager [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Terminating instance [ 869.641531] env[61985]: DEBUG nova.compute.manager [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 869.641949] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 869.643348] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c27d154-8a5a-4eb6-b5ab-8579b3ad7193 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.656913] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 869.657952] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77729e4d-7477-45fe-8497-256eee09a0b6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.666554] env[61985]: DEBUG oslo_vmware.api [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for the task: (returnval){ [ 869.666554] env[61985]: value = "task-935957" [ 869.666554] env[61985]: _type = "Task" [ 869.666554] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.678563] env[61985]: DEBUG oslo_vmware.api [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935957, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.741874] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.842717] env[61985]: DEBUG nova.compute.manager [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 869.873880] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.874400] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.874606] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.874818] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.874978] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.875161] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.875388] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.875557] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.875738] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.876342] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.876412] env[61985]: DEBUG nova.virt.hardware [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.877462] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a72c1d-45a6-426d-9cfc-b190aa55c466 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.887799] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24524f39-e3b6-4e27-811b-786b24eee8ed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.940032] env[61985]: DEBUG nova.scheduler.client.report [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.953977] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935956, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.327485} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.954546] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.955592] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb2337c-ac0e-4e6c-8cb4-f2cc91078620 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.985659] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 59e742d9-c51a-47dd-9ef2-2f070eae94c4/59e742d9-c51a-47dd-9ef2-2f070eae94c4.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.986428] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10c3dc3c-8864-4edf-b8fd-cb876f0c02a8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.008726] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 870.008726] env[61985]: value = "task-935958" [ 870.008726] env[61985]: _type = "Task" [ 870.008726] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.021537] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935958, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.179962] env[61985]: DEBUG oslo_vmware.api [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935957, 'name': PowerOffVM_Task, 'duration_secs': 0.244312} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.180339] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 870.180562] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 870.180901] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64cfc137-4701-414a-ae67-8dcb55e28650 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.310610] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 870.311480] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 870.311480] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Deleting the datastore file [datastore2] 17bdf8ec-13d2-459c-bc8e-db6a274fc27e {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.311480] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3c4039f-7309-4ed8-bf84-b3dee2cde97a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.319405] env[61985]: DEBUG oslo_vmware.api [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for the task: (returnval){ [ 870.319405] env[61985]: value = "task-935960" [ 870.319405] env[61985]: _type = "Task" [ 870.319405] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.329810] env[61985]: DEBUG oslo_vmware.api [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.443116] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.443826] env[61985]: DEBUG nova.compute.manager [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 870.446779] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 28.259s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.446947] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.447061] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 870.447308] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.810s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.448795] env[61985]: INFO nova.compute.claims [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.455021] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347e4459-a994-48e1-9f4e-e0602196b3d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.465052] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9440c307-a684-4702-bd99-821ecbe069b2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.483923] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c95fff-e3f9-4f12-8ec5-91a3c1c7e29f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.496459] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80747de4-fe9b-4745-8157-eab363c9fcdb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.529179] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180744MB free_disk=43GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 870.529179] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.537554] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935958, 'name': ReconfigVM_Task, 'duration_secs': 0.413962} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.537857] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 59e742d9-c51a-47dd-9ef2-2f070eae94c4/59e742d9-c51a-47dd-9ef2-2f070eae94c4.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 870.538511] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20415368-51e4-4f0b-82a5-d535c569b7a8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.548025] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 870.548025] env[61985]: value = "task-935961" [ 870.548025] env[61985]: _type = "Task" [ 870.548025] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.563900] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935961, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.830058] env[61985]: DEBUG oslo_vmware.api [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Task: {'id': task-935960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.303732} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.830345] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.830780] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 870.830780] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 870.830897] env[61985]: INFO nova.compute.manager [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Took 1.19 seconds to destroy the instance on the hypervisor. [ 870.831785] env[61985]: DEBUG oslo.service.loopingcall [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.831785] env[61985]: DEBUG nova.compute.manager [-] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.831785] env[61985]: DEBUG nova.network.neutron [-] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 870.954861] env[61985]: DEBUG nova.compute.utils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 870.956457] env[61985]: DEBUG nova.compute.manager [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 870.956563] env[61985]: DEBUG nova.network.neutron [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 871.060595] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935961, 'name': Rename_Task, 'duration_secs': 0.179768} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.060996] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 871.061259] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99c7f52e-508c-4bbe-850c-6a664d16e293 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.070429] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 871.070429] env[61985]: value = "task-935962" [ 871.070429] env[61985]: _type = "Task" [ 871.070429] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.080988] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935962, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.082245] env[61985]: DEBUG nova.policy [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6c18c735204b45ac9ded50c8c7874e04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '006a357baa9f4bd288cbec53a0be88cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.210666] env[61985]: DEBUG nova.compute.manager [req-c1f6e4e3-29e5-4d1b-b473-4a619ee2d12e req-eb0b3eba-cdc1-4127-803c-6882f73ae7ac service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Received event network-vif-deleted-9669d21e-aa85-402c-81f8-b83cb1368351 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 871.211608] env[61985]: INFO nova.compute.manager [req-c1f6e4e3-29e5-4d1b-b473-4a619ee2d12e req-eb0b3eba-cdc1-4127-803c-6882f73ae7ac service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Neutron deleted interface 9669d21e-aa85-402c-81f8-b83cb1368351; detaching it from the instance and deleting it from the info cache [ 871.211608] env[61985]: DEBUG nova.network.neutron [req-c1f6e4e3-29e5-4d1b-b473-4a619ee2d12e req-eb0b3eba-cdc1-4127-803c-6882f73ae7ac service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.258838] env[61985]: DEBUG nova.compute.manager [req-83dadbc2-a6ea-4089-94f4-b2d11a62cb8b req-faca47cf-83ab-450c-9c28-a4be9820890d service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Received event network-vif-plugged-4d18bb14-4f82-46b7-b128-9acc3eaae43f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 871.259059] env[61985]: DEBUG oslo_concurrency.lockutils [req-83dadbc2-a6ea-4089-94f4-b2d11a62cb8b req-faca47cf-83ab-450c-9c28-a4be9820890d service nova] Acquiring lock "ce757dda-f58a-47b3-b319-e6b01f05b20b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.259314] env[61985]: DEBUG oslo_concurrency.lockutils [req-83dadbc2-a6ea-4089-94f4-b2d11a62cb8b req-faca47cf-83ab-450c-9c28-a4be9820890d service nova] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.259314] env[61985]: DEBUG oslo_concurrency.lockutils [req-83dadbc2-a6ea-4089-94f4-b2d11a62cb8b req-faca47cf-83ab-450c-9c28-a4be9820890d service nova] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.259461] env[61985]: DEBUG nova.compute.manager [req-83dadbc2-a6ea-4089-94f4-b2d11a62cb8b req-faca47cf-83ab-450c-9c28-a4be9820890d service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] No waiting events found dispatching network-vif-plugged-4d18bb14-4f82-46b7-b128-9acc3eaae43f {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 871.259625] env[61985]: WARNING nova.compute.manager [req-83dadbc2-a6ea-4089-94f4-b2d11a62cb8b req-faca47cf-83ab-450c-9c28-a4be9820890d service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Received unexpected event network-vif-plugged-4d18bb14-4f82-46b7-b128-9acc3eaae43f for instance with vm_state building and task_state spawning. [ 871.464118] env[61985]: DEBUG nova.compute.manager [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 871.577603] env[61985]: DEBUG nova.network.neutron [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Successfully updated port: 4d18bb14-4f82-46b7-b128-9acc3eaae43f {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.586624] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935962, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.684992] env[61985]: DEBUG nova.network.neutron [-] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.719398] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5fe51b27-cfed-45a4-859c-82fb0d2ac8f4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.736963] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b6b23c-cea9-49cc-ba7a-71bb6f4c3ce4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.753449] env[61985]: DEBUG nova.network.neutron [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Successfully created port: e25b97ce-77e2-472d-a7a2-42899e22aab2 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.783666] env[61985]: DEBUG nova.compute.manager [req-c1f6e4e3-29e5-4d1b-b473-4a619ee2d12e req-eb0b3eba-cdc1-4127-803c-6882f73ae7ac service nova] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Detach interface failed, port_id=9669d21e-aa85-402c-81f8-b83cb1368351, reason: Instance 17bdf8ec-13d2-459c-bc8e-db6a274fc27e could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 871.955243] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f6918c8-dc2b-4597-aafa-a5d8cbc77aa7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.963791] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5197ac-0dd3-4f57-be36-387c3cafad70 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.007900] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87df3ca5-2f45-437e-b038-3b208230c234 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.021205] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc65a41f-129a-4b9c-8432-1d6d7b8a2f7d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.038181] env[61985]: DEBUG nova.compute.provider_tree [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.083246] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "refresh_cache-ce757dda-f58a-47b3-b319-e6b01f05b20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.083408] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "refresh_cache-ce757dda-f58a-47b3-b319-e6b01f05b20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.083545] env[61985]: DEBUG nova.network.neutron [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 872.084718] env[61985]: DEBUG oslo_vmware.api [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-935962, 'name': PowerOnVM_Task, 'duration_secs': 0.960795} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.084908] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 872.085128] env[61985]: INFO nova.compute.manager [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Took 9.28 seconds to spawn the instance on the hypervisor. [ 872.085317] env[61985]: DEBUG nova.compute.manager [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.086117] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73b8d8a-1a71-4928-85ac-8944fb020c0d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.189918] env[61985]: INFO nova.compute.manager [-] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Took 1.36 seconds to deallocate network for instance. [ 872.510450] env[61985]: DEBUG nova.compute.manager [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 872.544103] env[61985]: DEBUG nova.scheduler.client.report [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.549428] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.553022] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.553022] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.553022] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.553022] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.553022] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.553274] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.553274] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.553274] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.553274] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.553274] env[61985]: DEBUG nova.virt.hardware [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.553411] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c464c8-0906-4988-bef0-b10817d0f64b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.564098] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77de03ef-f071-4ccf-b8b1-e33cf8bd2d12 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.608334] env[61985]: INFO nova.compute.manager [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Took 49.99 seconds to build instance. [ 872.661029] env[61985]: DEBUG nova.network.neutron [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 872.698509] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.965747] env[61985]: DEBUG nova.network.neutron [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Updating instance_info_cache with network_info: [{"id": "4d18bb14-4f82-46b7-b128-9acc3eaae43f", "address": "fa:16:3e:62:e1:57", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d18bb14-4f", "ovs_interfaceid": "4d18bb14-4f82-46b7-b128-9acc3eaae43f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.058370] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.611s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.059962] env[61985]: DEBUG nova.compute.manager [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 873.063661] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.500s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.065842] env[61985]: INFO nova.compute.claims [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.110219] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8565cbe6-695d-43f8-85c1-d90aa6c5df88 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.565s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.471568] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "refresh_cache-ce757dda-f58a-47b3-b319-e6b01f05b20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.471929] env[61985]: DEBUG nova.compute.manager [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Instance network_info: |[{"id": "4d18bb14-4f82-46b7-b128-9acc3eaae43f", "address": "fa:16:3e:62:e1:57", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d18bb14-4f", "ovs_interfaceid": "4d18bb14-4f82-46b7-b128-9acc3eaae43f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 873.472385] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:e1:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd829efb7-e98e-4b67-bd03-b0888287dbfd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4d18bb14-4f82-46b7-b128-9acc3eaae43f', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.483360] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Creating folder: Project (4923009773c047519b883745c59e3516). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 873.483781] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d0322c2-177e-4c26-8f95-8b9ea35df8b7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.497838] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Created folder: Project (4923009773c047519b883745c59e3516) in parent group-v211285. [ 873.498103] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Creating folder: Instances. Parent ref: group-v211379. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 873.498366] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69a14dc1-4150-48c4-af7a-0cbebe881320 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.510198] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Created folder: Instances in parent group-v211379. [ 873.510536] env[61985]: DEBUG oslo.service.loopingcall [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.513999] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 873.513999] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea2892d8-eac3-425b-81e9-f5dcc45a6335 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.536438] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.536438] env[61985]: value = "task-935965" [ 873.536438] env[61985]: _type = "Task" [ 873.536438] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.546549] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935965, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.572359] env[61985]: DEBUG nova.compute.utils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.579547] env[61985]: DEBUG nova.compute.manager [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 873.579738] env[61985]: DEBUG nova.network.neutron [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 873.615057] env[61985]: DEBUG nova.compute.manager [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.641322] env[61985]: DEBUG nova.network.neutron [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Successfully updated port: e25b97ce-77e2-472d-a7a2-42899e22aab2 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.674023] env[61985]: DEBUG nova.policy [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'baf9793f2d8a41698e11196f639fd749', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd49f3925a4074d779d6dd68164545cfb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.747110] env[61985]: DEBUG nova.compute.manager [req-5ace9f09-82e2-431d-b918-589c5ddb5e1d req-8b09b5f3-7238-4b8e-81aa-7f49832582c3 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Received event network-changed-4d18bb14-4f82-46b7-b128-9acc3eaae43f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 873.747298] env[61985]: DEBUG nova.compute.manager [req-5ace9f09-82e2-431d-b918-589c5ddb5e1d req-8b09b5f3-7238-4b8e-81aa-7f49832582c3 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Refreshing instance network info cache due to event network-changed-4d18bb14-4f82-46b7-b128-9acc3eaae43f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 873.747710] env[61985]: DEBUG oslo_concurrency.lockutils [req-5ace9f09-82e2-431d-b918-589c5ddb5e1d req-8b09b5f3-7238-4b8e-81aa-7f49832582c3 service nova] Acquiring lock "refresh_cache-ce757dda-f58a-47b3-b319-e6b01f05b20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.748033] env[61985]: DEBUG oslo_concurrency.lockutils [req-5ace9f09-82e2-431d-b918-589c5ddb5e1d req-8b09b5f3-7238-4b8e-81aa-7f49832582c3 service nova] Acquired lock "refresh_cache-ce757dda-f58a-47b3-b319-e6b01f05b20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.748749] env[61985]: DEBUG nova.network.neutron [req-5ace9f09-82e2-431d-b918-589c5ddb5e1d req-8b09b5f3-7238-4b8e-81aa-7f49832582c3 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Refreshing network info cache for port 4d18bb14-4f82-46b7-b128-9acc3eaae43f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 874.056737] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935965, 'name': CreateVM_Task, 'duration_secs': 0.380955} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.057018] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 874.058057] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.058384] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.058845] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 874.059251] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35256430-6049-45e3-9fc1-e787ae0ccff2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.065064] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 874.065064] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e0ad3d-03ba-c5c7-7258-b437541d7c2f" [ 874.065064] env[61985]: _type = "Task" [ 874.065064] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.075820] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e0ad3d-03ba-c5c7-7258-b437541d7c2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.084710] env[61985]: DEBUG nova.compute.manager [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 874.134888] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquiring lock "refresh_cache-5b3a8ba3-48cb-472a-ab45-4e51af68812c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.134888] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquired lock "refresh_cache-5b3a8ba3-48cb-472a-ab45-4e51af68812c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.135066] env[61985]: DEBUG nova.network.neutron [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.143833] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.186825] env[61985]: DEBUG nova.network.neutron [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Successfully created port: 885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.539900] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3f8acc-d25c-44bb-80d4-f874005ab43d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.549089] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1b4422-5b98-4e86-871a-61730869c098 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.587857] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b85048-9961-44fe-bf6f-7da053de5cbb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.596532] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "87006fbf-ea90-4d9a-88af-001de424ac14" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.596786] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "87006fbf-ea90-4d9a-88af-001de424ac14" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.604884] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e0ad3d-03ba-c5c7-7258-b437541d7c2f, 'name': SearchDatastore_Task, 'duration_secs': 0.015249} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.606878] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.607153] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.607460] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.607649] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.607922] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.608298] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a5abc1d-8d74-4ac2-aeef-80fc8506d80a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.611929] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28676256-fee4-4e10-abba-5a6dadcdcb0d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.631012] env[61985]: DEBUG nova.compute.provider_tree [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.634022] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.634022] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 874.634991] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7328a491-85e0-4faa-897e-18bfe32074af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.647024] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 874.647024] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52be1bc6-af52-8442-9d8f-22a96bbd161a" [ 874.647024] env[61985]: _type = "Task" [ 874.647024] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.654317] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52be1bc6-af52-8442-9d8f-22a96bbd161a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.686461] env[61985]: DEBUG nova.network.neutron [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 874.853242] env[61985]: DEBUG nova.network.neutron [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Updating instance_info_cache with network_info: [{"id": "e25b97ce-77e2-472d-a7a2-42899e22aab2", "address": "fa:16:3e:ea:45:2a", "network": {"id": "8d56ab44-9a85-46d5-9ce5-9bfe7830ba5f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1009275674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "006a357baa9f4bd288cbec53a0be88cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape25b97ce-77", "ovs_interfaceid": "e25b97ce-77e2-472d-a7a2-42899e22aab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.099345] env[61985]: DEBUG nova.network.neutron [req-5ace9f09-82e2-431d-b918-589c5ddb5e1d req-8b09b5f3-7238-4b8e-81aa-7f49832582c3 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Updated VIF entry in instance network info cache for port 4d18bb14-4f82-46b7-b128-9acc3eaae43f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 875.100027] env[61985]: DEBUG nova.network.neutron [req-5ace9f09-82e2-431d-b918-589c5ddb5e1d req-8b09b5f3-7238-4b8e-81aa-7f49832582c3 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Updating instance_info_cache with network_info: [{"id": "4d18bb14-4f82-46b7-b128-9acc3eaae43f", "address": "fa:16:3e:62:e1:57", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d18bb14-4f", "ovs_interfaceid": "4d18bb14-4f82-46b7-b128-9acc3eaae43f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.105415] env[61985]: DEBUG nova.compute.manager [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 875.133641] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.133896] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.134074] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.134271] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.134426] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.134584] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.134806] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.134973] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.135265] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.135480] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.135728] env[61985]: DEBUG nova.virt.hardware [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.136853] env[61985]: DEBUG nova.scheduler.client.report [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.143323] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28eb10e8-7b4e-426a-b64b-5bddc90ac765 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.152452] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c163dbf5-99ab-4e11-82a2-240b185ad690 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.161886] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52be1bc6-af52-8442-9d8f-22a96bbd161a, 'name': SearchDatastore_Task, 'duration_secs': 0.014888} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.162750] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a0daba2-fb3f-4aae-b534-4766a815b955 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.177186] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 875.177186] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522d6adb-0229-0390-1e32-eaa77e5cc044" [ 875.177186] env[61985]: _type = "Task" [ 875.177186] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.188021] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522d6adb-0229-0390-1e32-eaa77e5cc044, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.356446] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Releasing lock "refresh_cache-5b3a8ba3-48cb-472a-ab45-4e51af68812c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.358030] env[61985]: DEBUG nova.compute.manager [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Instance network_info: |[{"id": "e25b97ce-77e2-472d-a7a2-42899e22aab2", "address": "fa:16:3e:ea:45:2a", "network": {"id": "8d56ab44-9a85-46d5-9ce5-9bfe7830ba5f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1009275674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "006a357baa9f4bd288cbec53a0be88cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape25b97ce-77", "ovs_interfaceid": "e25b97ce-77e2-472d-a7a2-42899e22aab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.358204] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:45:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ff90ec9-3c7e-4e76-b409-fcf37fc588d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e25b97ce-77e2-472d-a7a2-42899e22aab2', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.367318] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Creating folder: Project (006a357baa9f4bd288cbec53a0be88cc). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 875.367689] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22d7d499-adf7-46dd-be5d-fb5c66efc39f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.384747] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Created folder: Project (006a357baa9f4bd288cbec53a0be88cc) in parent group-v211285. [ 875.385181] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Creating folder: Instances. Parent ref: group-v211382. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 875.385341] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d02cca8a-974c-486b-abde-0fea619599c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.400318] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Created folder: Instances in parent group-v211382. [ 875.400659] env[61985]: DEBUG oslo.service.loopingcall [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.400913] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 875.401187] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab0cff54-aea1-416b-905f-178ce0d7c48a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.425186] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.425186] env[61985]: value = "task-935968" [ 875.425186] env[61985]: _type = "Task" [ 875.425186] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.434793] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935968, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.607855] env[61985]: DEBUG oslo_concurrency.lockutils [req-5ace9f09-82e2-431d-b918-589c5ddb5e1d req-8b09b5f3-7238-4b8e-81aa-7f49832582c3 service nova] Releasing lock "refresh_cache-ce757dda-f58a-47b3-b319-e6b01f05b20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.644773] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.581s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.645513] env[61985]: DEBUG nova.compute.manager [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 875.648148] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.769s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.648387] env[61985]: DEBUG nova.objects.instance [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lazy-loading 'resources' on Instance uuid 402ef006-2835-4239-a3c4-1c37a7bba9a5 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.690348] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522d6adb-0229-0390-1e32-eaa77e5cc044, 'name': SearchDatastore_Task, 'duration_secs': 0.010468} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.690652] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.690917] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] ce757dda-f58a-47b3-b319-e6b01f05b20b/ce757dda-f58a-47b3-b319-e6b01f05b20b.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 875.691449] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2e88e4f-40ee-4ca1-832a-4c0fb634d48a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.704059] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 875.704059] env[61985]: value = "task-935969" [ 875.704059] env[61985]: _type = "Task" [ 875.704059] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.713357] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935969, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.862798] env[61985]: DEBUG nova.compute.manager [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Received event network-vif-plugged-e25b97ce-77e2-472d-a7a2-42899e22aab2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 875.863096] env[61985]: DEBUG oslo_concurrency.lockutils [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] Acquiring lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.863466] env[61985]: DEBUG oslo_concurrency.lockutils [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] Lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.863688] env[61985]: DEBUG oslo_concurrency.lockutils [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] Lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.863863] env[61985]: DEBUG nova.compute.manager [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] No waiting events found dispatching network-vif-plugged-e25b97ce-77e2-472d-a7a2-42899e22aab2 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.864108] env[61985]: WARNING nova.compute.manager [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Received unexpected event network-vif-plugged-e25b97ce-77e2-472d-a7a2-42899e22aab2 for instance with vm_state building and task_state spawning. [ 875.864394] env[61985]: DEBUG nova.compute.manager [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Received event network-changed-e25b97ce-77e2-472d-a7a2-42899e22aab2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 875.864574] env[61985]: DEBUG nova.compute.manager [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Refreshing instance network info cache due to event network-changed-e25b97ce-77e2-472d-a7a2-42899e22aab2. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 875.864818] env[61985]: DEBUG oslo_concurrency.lockutils [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] Acquiring lock "refresh_cache-5b3a8ba3-48cb-472a-ab45-4e51af68812c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.865091] env[61985]: DEBUG oslo_concurrency.lockutils [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] Acquired lock "refresh_cache-5b3a8ba3-48cb-472a-ab45-4e51af68812c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.865262] env[61985]: DEBUG nova.network.neutron [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Refreshing network info cache for port e25b97ce-77e2-472d-a7a2-42899e22aab2 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 875.945755] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.946081] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.946229] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935968, 'name': CreateVM_Task, 'duration_secs': 0.477313} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.946443] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 875.947485] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.947485] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.947706] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.948670] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa3eaecc-7e2c-4279-b520-df16ff9f0e1f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.955316] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 875.955316] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52082eeb-1dd7-7285-3e62-002620596af6" [ 875.955316] env[61985]: _type = "Task" [ 875.955316] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.966103] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52082eeb-1dd7-7285-3e62-002620596af6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.156225] env[61985]: DEBUG nova.compute.utils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.159025] env[61985]: DEBUG nova.compute.manager [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.159383] env[61985]: DEBUG nova.network.neutron [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 876.212382] env[61985]: DEBUG nova.policy [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '344f7f484f4240cd9dc32b200b697e12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfc57e6393c94545be32165d41230db3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 876.219981] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935969, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48732} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.220295] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] ce757dda-f58a-47b3-b319-e6b01f05b20b/ce757dda-f58a-47b3-b319-e6b01f05b20b.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 876.220543] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 876.220914] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0229c2cb-99ba-4042-a784-ddfd23db56bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.228959] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 876.228959] env[61985]: value = "task-935970" [ 876.228959] env[61985]: _type = "Task" [ 876.228959] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.238474] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935970, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.472202] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52082eeb-1dd7-7285-3e62-002620596af6, 'name': SearchDatastore_Task, 'duration_secs': 0.061176} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.472202] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.472202] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.472330] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.472438] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.472630] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.472905] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1046e79-6004-4c0d-b002-ccb7efbd42c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.487241] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.487465] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 876.488286] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5939878d-434b-4c01-b039-7102c03d509f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.497506] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 876.497506] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a3a891-c2f2-5e8a-576d-b908ac1b2a39" [ 876.497506] env[61985]: _type = "Task" [ 876.497506] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.509773] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a3a891-c2f2-5e8a-576d-b908ac1b2a39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.581713] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351b7dbe-b734-4570-bf81-10f4cbdc9f9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.592224] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bc8994-ca6a-40ce-8511-9fa6126d621b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.629992] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277f6e88-84fc-464d-abdb-ee80ec8dda28 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.633498] env[61985]: DEBUG nova.network.neutron [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Updated VIF entry in instance network info cache for port e25b97ce-77e2-472d-a7a2-42899e22aab2. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 876.633931] env[61985]: DEBUG nova.network.neutron [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Updating instance_info_cache with network_info: [{"id": "e25b97ce-77e2-472d-a7a2-42899e22aab2", "address": "fa:16:3e:ea:45:2a", "network": {"id": "8d56ab44-9a85-46d5-9ce5-9bfe7830ba5f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1009275674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "006a357baa9f4bd288cbec53a0be88cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape25b97ce-77", "ovs_interfaceid": "e25b97ce-77e2-472d-a7a2-42899e22aab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.642349] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b13221-2130-4b9c-901b-5311f9d10a93 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.662299] env[61985]: DEBUG nova.compute.provider_tree [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.664292] env[61985]: DEBUG nova.compute.manager [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 876.739728] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935970, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067613} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.739885] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.740693] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95493ee8-cf15-4f54-8d76-37625761921d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.763826] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] ce757dda-f58a-47b3-b319-e6b01f05b20b/ce757dda-f58a-47b3-b319-e6b01f05b20b.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.764220] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cffd6baa-033e-4589-a259-cee5310e4e19 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.787314] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 876.787314] env[61985]: value = "task-935971" [ 876.787314] env[61985]: _type = "Task" [ 876.787314] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.795201] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935971, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.908072] env[61985]: DEBUG nova.network.neutron [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Successfully created port: 5b4f0fea-8589-4efc-b9b3-b59de5621860 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.965190] env[61985]: DEBUG nova.network.neutron [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Successfully updated port: 885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.015574] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a3a891-c2f2-5e8a-576d-b908ac1b2a39, 'name': SearchDatastore_Task, 'duration_secs': 0.024432} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.016433] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2bb449d-1928-45b0-9e2c-83d5cee092fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.022995] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 877.022995] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f15399-76bd-23ba-e90d-23eb66f6a8fc" [ 877.022995] env[61985]: _type = "Task" [ 877.022995] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.034849] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f15399-76bd-23ba-e90d-23eb66f6a8fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.136872] env[61985]: DEBUG oslo_concurrency.lockutils [req-f7730ba3-4d90-4053-9353-8fb9e90594ad req-82dcc933-05bf-4982-a434-1169f2ea07ae service nova] Releasing lock "refresh_cache-5b3a8ba3-48cb-472a-ab45-4e51af68812c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.168315] env[61985]: DEBUG nova.scheduler.client.report [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.296625] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935971, 'name': ReconfigVM_Task, 'duration_secs': 0.32279} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.296968] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Reconfigured VM instance instance-0000003d to attach disk [datastore1] ce757dda-f58a-47b3-b319-e6b01f05b20b/ce757dda-f58a-47b3-b319-e6b01f05b20b.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.297657] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0714d328-f1de-4cb5-87bb-200548abc5cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.305912] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 877.305912] env[61985]: value = "task-935972" [ 877.305912] env[61985]: _type = "Task" [ 877.305912] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.322104] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935972, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.467799] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.468130] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquired lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.468272] env[61985]: DEBUG nova.network.neutron [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.535873] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f15399-76bd-23ba-e90d-23eb66f6a8fc, 'name': SearchDatastore_Task, 'duration_secs': 0.01153} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.536218] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.536578] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 5b3a8ba3-48cb-472a-ab45-4e51af68812c/5b3a8ba3-48cb-472a-ab45-4e51af68812c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 877.536907] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad4af7f2-2cba-4939-8368-a6be4b7afce2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.544825] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 877.544825] env[61985]: value = "task-935973" [ 877.544825] env[61985]: _type = "Task" [ 877.544825] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.554772] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.677102] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.028s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.680084] env[61985]: DEBUG nova.compute.manager [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 877.682411] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.621s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.683992] env[61985]: INFO nova.compute.claims [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.704612] env[61985]: INFO nova.scheduler.client.report [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Deleted allocations for instance 402ef006-2835-4239-a3c4-1c37a7bba9a5 [ 877.708720] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.708873] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.709068] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.709320] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.709454] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.709626] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.709877] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.710072] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.710290] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.710484] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.710674] env[61985]: DEBUG nova.virt.hardware [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.711623] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051b0a8f-045b-4d97-9957-77993aefccbd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.723822] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7676f5dd-7167-4667-8110-01e83824706f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.819614] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935972, 'name': Rename_Task, 'duration_secs': 0.179058} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.819614] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 877.819614] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1c7847b-ff89-4ba7-9954-883098a282b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.827733] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 877.827733] env[61985]: value = "task-935974" [ 877.827733] env[61985]: _type = "Task" [ 877.827733] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.837967] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935974, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.877613] env[61985]: DEBUG oslo_vmware.rw_handles [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52071d0e-6b7c-83f0-a4c5-08f063edab51/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 877.881148] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f729e4b8-a893-407c-abe9-6f6e1c147909 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.888020] env[61985]: DEBUG oslo_vmware.rw_handles [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52071d0e-6b7c-83f0-a4c5-08f063edab51/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 877.888346] env[61985]: ERROR oslo_vmware.rw_handles [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52071d0e-6b7c-83f0-a4c5-08f063edab51/disk-0.vmdk due to incomplete transfer. [ 877.888649] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3ab8a14d-1d28-4e4e-8e7c-1331099fdd41 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.902373] env[61985]: DEBUG oslo_vmware.rw_handles [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52071d0e-6b7c-83f0-a4c5-08f063edab51/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 877.902373] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Uploaded image 9872176a-7a1f-42a1-895a-258479200089 to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 877.904723] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 877.907097] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-cac473d5-5356-430d-ac9b-525572a7dc83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.911006] env[61985]: DEBUG nova.compute.manager [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Received event network-vif-plugged-885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 877.911006] env[61985]: DEBUG oslo_concurrency.lockutils [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] Acquiring lock "c64056d8-0c22-48db-bf4a-a60f468bdad0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.911232] env[61985]: DEBUG oslo_concurrency.lockutils [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.911466] env[61985]: DEBUG oslo_concurrency.lockutils [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.911648] env[61985]: DEBUG nova.compute.manager [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] No waiting events found dispatching network-vif-plugged-885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.911825] env[61985]: WARNING nova.compute.manager [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Received unexpected event network-vif-plugged-885cda89-b90d-499a-a3b1-7155c07a1ce7 for instance with vm_state building and task_state spawning. [ 877.911993] env[61985]: DEBUG nova.compute.manager [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Received event network-changed-885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 877.912174] env[61985]: DEBUG nova.compute.manager [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Refreshing instance network info cache due to event network-changed-885cda89-b90d-499a-a3b1-7155c07a1ce7. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 877.912346] env[61985]: DEBUG oslo_concurrency.lockutils [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] Acquiring lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.922957] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 877.922957] env[61985]: value = "task-935975" [ 877.922957] env[61985]: _type = "Task" [ 877.922957] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.952813] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935975, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.027057] env[61985]: DEBUG nova.network.neutron [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 878.063409] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935973, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.221483] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d40f5aa-38b6-4b08-b7b4-10e9c43127b4 tempest-ServerDiagnosticsTest-1637129757 tempest-ServerDiagnosticsTest-1637129757-project-member] Lock "402ef006-2835-4239-a3c4-1c37a7bba9a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.361s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.245466] env[61985]: DEBUG nova.network.neutron [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Updating instance_info_cache with network_info: [{"id": "885cda89-b90d-499a-a3b1-7155c07a1ce7", "address": "fa:16:3e:7d:a3:a3", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap885cda89-b9", "ovs_interfaceid": "885cda89-b90d-499a-a3b1-7155c07a1ce7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.339465] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935974, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.435686] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935975, 'name': Destroy_Task} progress is 33%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.558728] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547795} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.558728] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 5b3a8ba3-48cb-472a-ab45-4e51af68812c/5b3a8ba3-48cb-472a-ab45-4e51af68812c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 878.558728] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.558728] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c0c24617-1533-4753-95c7-ebe80ffb2135 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.567153] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 878.567153] env[61985]: value = "task-935976" [ 878.567153] env[61985]: _type = "Task" [ 878.567153] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.577057] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935976, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.748889] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Releasing lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.748889] env[61985]: DEBUG nova.compute.manager [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance network_info: |[{"id": "885cda89-b90d-499a-a3b1-7155c07a1ce7", "address": "fa:16:3e:7d:a3:a3", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap885cda89-b9", "ovs_interfaceid": "885cda89-b90d-499a-a3b1-7155c07a1ce7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.749172] env[61985]: DEBUG oslo_concurrency.lockutils [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] Acquired lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.749366] env[61985]: DEBUG nova.network.neutron [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Refreshing network info cache for port 885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.750639] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:a3:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '885cda89-b90d-499a-a3b1-7155c07a1ce7', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.763207] env[61985]: DEBUG oslo.service.loopingcall [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.768939] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 878.771484] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3fab9931-890c-4491-aa38-99be903f19cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.801752] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.801752] env[61985]: value = "task-935977" [ 878.801752] env[61985]: _type = "Task" [ 878.801752] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.811423] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935977, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.840363] env[61985]: DEBUG oslo_vmware.api [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-935974, 'name': PowerOnVM_Task, 'duration_secs': 0.5799} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.840662] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 878.840871] env[61985]: INFO nova.compute.manager [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Took 9.00 seconds to spawn the instance on the hypervisor. [ 878.841096] env[61985]: DEBUG nova.compute.manager [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.841949] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282519c7-3212-469e-807b-c4e02952e7a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.857860] env[61985]: DEBUG nova.network.neutron [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Successfully updated port: 5b4f0fea-8589-4efc-b9b3-b59de5621860 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.936882] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935975, 'name': Destroy_Task, 'duration_secs': 0.853952} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.937188] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Destroyed the VM [ 878.937463] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 878.937746] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b4f9d4cd-f636-4faa-b148-9d70ce85a773 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.951486] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 878.951486] env[61985]: value = "task-935978" [ 878.951486] env[61985]: _type = "Task" [ 878.951486] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.968892] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935978, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.078881] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935976, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.187229} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.079142] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.080032] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9880ab7f-a25b-46f5-a6e6-f9b422a56a98 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.110135] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 5b3a8ba3-48cb-472a-ab45-4e51af68812c/5b3a8ba3-48cb-472a-ab45-4e51af68812c.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.113269] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f7fc90a-83b6-4155-b72f-88d0bbf7bbcb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.138651] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 879.138651] env[61985]: value = "task-935979" [ 879.138651] env[61985]: _type = "Task" [ 879.138651] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.152645] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935979, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.184540] env[61985]: DEBUG nova.network.neutron [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Updated VIF entry in instance network info cache for port 885cda89-b90d-499a-a3b1-7155c07a1ce7. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 879.185178] env[61985]: DEBUG nova.network.neutron [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Updating instance_info_cache with network_info: [{"id": "885cda89-b90d-499a-a3b1-7155c07a1ce7", "address": "fa:16:3e:7d:a3:a3", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap885cda89-b9", "ovs_interfaceid": "885cda89-b90d-499a-a3b1-7155c07a1ce7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.217714] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c551114d-40d6-448d-b88b-647036ddc27b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.226859] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d587300e-c954-4bd8-9c4e-f85a980044a4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.263227] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f5224f-4bcb-426d-9343-62f3fe07ef33 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.272818] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e5a359-79f8-4e4a-8519-1769c3067fe5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.288150] env[61985]: DEBUG nova.compute.provider_tree [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.315552] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935977, 'name': CreateVM_Task, 'duration_secs': 0.420736} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.315552] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 879.315552] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.315866] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.315943] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.316697] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df0605c6-a40f-4673-b9c3-5d26dc9fe18b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.323533] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 879.323533] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cb08ac-270f-e6df-17a7-e9e11150b724" [ 879.323533] env[61985]: _type = "Task" [ 879.323533] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.330544] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cb08ac-270f-e6df-17a7-e9e11150b724, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.362841] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.362841] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.362841] env[61985]: DEBUG nova.network.neutron [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.367854] env[61985]: INFO nova.compute.manager [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Took 41.09 seconds to build instance. [ 879.465080] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935978, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.654019] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935979, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.691696] env[61985]: DEBUG oslo_concurrency.lockutils [req-18e0bbf0-c01b-4533-825b-e151eb73dc16 req-c3e90638-fca2-4719-83dd-0bae9ee07afb service nova] Releasing lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.792693] env[61985]: DEBUG nova.scheduler.client.report [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.833951] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cb08ac-270f-e6df-17a7-e9e11150b724, 'name': SearchDatastore_Task, 'duration_secs': 0.010286} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.834297] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.834578] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.834845] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.835026] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.835246] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.835624] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8888516-9a79-477d-b8f0-197c6b76abdc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.845311] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.845518] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 879.846281] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07f11ddd-9447-4f11-9ec3-a2ac11a52044 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.852199] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 879.852199] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525dbf29-c607-c156-9560-d36e7199e071" [ 879.852199] env[61985]: _type = "Task" [ 879.852199] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.860266] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525dbf29-c607-c156-9560-d36e7199e071, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.871626] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7c8239b4-7961-4293-80b4-1bbea7157df4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.466s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.966203] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935978, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.022204] env[61985]: DEBUG nova.network.neutron [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 880.150467] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935979, 'name': ReconfigVM_Task, 'duration_secs': 0.700069} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.150764] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 5b3a8ba3-48cb-472a-ab45-4e51af68812c/5b3a8ba3-48cb-472a-ab45-4e51af68812c.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.151404] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1abf002a-0705-4c00-b72f-d73e26b936ef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.158438] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 880.158438] env[61985]: value = "task-935980" [ 880.158438] env[61985]: _type = "Task" [ 880.158438] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.167029] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935980, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.302551] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.303153] env[61985]: DEBUG nova.compute.manager [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 880.315246] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.664s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.316775] env[61985]: INFO nova.compute.claims [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.341017] env[61985]: DEBUG nova.compute.manager [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Received event network-vif-plugged-5b4f0fea-8589-4efc-b9b3-b59de5621860 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 880.341017] env[61985]: DEBUG oslo_concurrency.lockutils [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] Acquiring lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.341017] env[61985]: DEBUG oslo_concurrency.lockutils [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] Lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.341017] env[61985]: DEBUG oslo_concurrency.lockutils [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] Lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.341017] env[61985]: DEBUG nova.compute.manager [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] No waiting events found dispatching network-vif-plugged-5b4f0fea-8589-4efc-b9b3-b59de5621860 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.341343] env[61985]: WARNING nova.compute.manager [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Received unexpected event network-vif-plugged-5b4f0fea-8589-4efc-b9b3-b59de5621860 for instance with vm_state building and task_state spawning. [ 880.341343] env[61985]: DEBUG nova.compute.manager [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Received event network-changed-5b4f0fea-8589-4efc-b9b3-b59de5621860 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 880.341343] env[61985]: DEBUG nova.compute.manager [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Refreshing instance network info cache due to event network-changed-5b4f0fea-8589-4efc-b9b3-b59de5621860. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 880.341343] env[61985]: DEBUG oslo_concurrency.lockutils [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] Acquiring lock "refresh_cache-e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.366124] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525dbf29-c607-c156-9560-d36e7199e071, 'name': SearchDatastore_Task, 'duration_secs': 0.009583} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.368090] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a66d8bc-87a5-413d-bdde-7537afd27027 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.375071] env[61985]: DEBUG nova.compute.manager [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.378131] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 880.378131] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d2557f-b7d3-2184-1b5b-254c7ad5a514" [ 880.378131] env[61985]: _type = "Task" [ 880.378131] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.395020] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d2557f-b7d3-2184-1b5b-254c7ad5a514, 'name': SearchDatastore_Task, 'duration_secs': 0.010688} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.395020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.395020] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] c64056d8-0c22-48db-bf4a-a60f468bdad0/c64056d8-0c22-48db-bf4a-a60f468bdad0.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 880.395020] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09cc52b8-261f-4d29-ae85-46e267884bd6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.406497] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 880.406497] env[61985]: value = "task-935981" [ 880.406497] env[61985]: _type = "Task" [ 880.406497] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.417434] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.424737] env[61985]: DEBUG nova.network.neutron [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Updating instance_info_cache with network_info: [{"id": "5b4f0fea-8589-4efc-b9b3-b59de5621860", "address": "fa:16:3e:63:ba:48", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b4f0fea-85", "ovs_interfaceid": "5b4f0fea-8589-4efc-b9b3-b59de5621860", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.464487] env[61985]: DEBUG oslo_vmware.api [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-935978, 'name': RemoveSnapshot_Task, 'duration_secs': 1.031742} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.464770] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 880.465014] env[61985]: INFO nova.compute.manager [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Took 17.77 seconds to snapshot the instance on the hypervisor. [ 880.670030] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935980, 'name': Rename_Task, 'duration_secs': 0.273734} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.670399] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 880.670715] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-feee6d20-533a-446e-bb8c-3698cf10481a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.679125] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 880.679125] env[61985]: value = "task-935982" [ 880.679125] env[61985]: _type = "Task" [ 880.679125] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.691027] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935982, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.826335] env[61985]: DEBUG nova.compute.utils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.830810] env[61985]: DEBUG nova.compute.manager [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 880.831085] env[61985]: DEBUG nova.network.neutron [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 880.900604] env[61985]: DEBUG nova.policy [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2d6440c03234eaea34bb76547507d3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4ebaf2c61d940d3bab278994b54b418', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.919415] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935981, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.921120] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.927232] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.927703] env[61985]: DEBUG nova.compute.manager [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Instance network_info: |[{"id": "5b4f0fea-8589-4efc-b9b3-b59de5621860", "address": "fa:16:3e:63:ba:48", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b4f0fea-85", "ovs_interfaceid": "5b4f0fea-8589-4efc-b9b3-b59de5621860", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.928114] env[61985]: DEBUG oslo_concurrency.lockutils [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] Acquired lock "refresh_cache-e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.928383] env[61985]: DEBUG nova.network.neutron [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Refreshing network info cache for port 5b4f0fea-8589-4efc-b9b3-b59de5621860 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 880.930765] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:ba:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bff6c3a1-cc80-46ca-86c0-6dbb029edddb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5b4f0fea-8589-4efc-b9b3-b59de5621860', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.939124] env[61985]: DEBUG oslo.service.loopingcall [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.939708] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 880.939983] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ffb733b-8423-4713-b2df-14c20d3b920b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.963854] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.963854] env[61985]: value = "task-935983" [ 880.963854] env[61985]: _type = "Task" [ 880.963854] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.977506] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935983, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.051249] env[61985]: DEBUG nova.compute.manager [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Found 3 images (rotation: 2) {{(pid=61985) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4560}} [ 881.051249] env[61985]: DEBUG nova.compute.manager [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Rotating out 1 backups {{(pid=61985) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4568}} [ 881.051249] env[61985]: DEBUG nova.compute.manager [None req-a132d5cf-b5ce-4a9b-bb8a-7e238a996037 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Deleting image fc3d508b-a533-4ace-90cf-e4f991a60e00 {{(pid=61985) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4573}} [ 881.190087] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935982, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.331953] env[61985]: DEBUG nova.compute.manager [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 881.354665] env[61985]: DEBUG nova.network.neutron [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Successfully created port: cfee741d-7a1d-4e4f-8432-1603248a1ba4 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.415779] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935981, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.71935} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.415779] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] c64056d8-0c22-48db-bf4a-a60f468bdad0/c64056d8-0c22-48db-bf4a-a60f468bdad0.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 881.415779] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.416012] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d77b0f60-e098-4ca7-a18d-83d537af62b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.424970] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 881.424970] env[61985]: value = "task-935984" [ 881.424970] env[61985]: _type = "Task" [ 881.424970] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.434690] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935984, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.474626] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935983, 'name': CreateVM_Task, 'duration_secs': 0.407023} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.477919] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 881.478770] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.478949] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.480377] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.480649] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0941bb78-62f1-493c-9307-f326fa64a9c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.485595] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 881.485595] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524fd153-725c-f75e-9d67-c287289ee548" [ 881.485595] env[61985]: _type = "Task" [ 881.485595] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.495264] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524fd153-725c-f75e-9d67-c287289ee548, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.510745] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae92469e-2a71-479e-9756-9d24ad08e02f tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.511015] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae92469e-2a71-479e-9756-9d24ad08e02f tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.511620] env[61985]: DEBUG nova.objects.instance [None req-ae92469e-2a71-479e-9756-9d24ad08e02f tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'flavor' on Instance uuid 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.696202] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935982, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.775169] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa392d6-0b62-4ba6-9a57-d5ab95cb6817 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.782079] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d52d83d-bced-49f4-947e-20496ed7ac0f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.818832] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f3c94d-2919-4a05-8990-ad561a0f4449 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.827278] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e7653d-5683-40c5-a103-f3e320a5fbc0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.844402] env[61985]: DEBUG nova.compute.provider_tree [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.891300] env[61985]: DEBUG nova.network.neutron [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Updated VIF entry in instance network info cache for port 5b4f0fea-8589-4efc-b9b3-b59de5621860. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 881.891794] env[61985]: DEBUG nova.network.neutron [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Updating instance_info_cache with network_info: [{"id": "5b4f0fea-8589-4efc-b9b3-b59de5621860", "address": "fa:16:3e:63:ba:48", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b4f0fea-85", "ovs_interfaceid": "5b4f0fea-8589-4efc-b9b3-b59de5621860", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.935498] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935984, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.172386} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.935994] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.936639] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0e0bb5-177d-4942-9c2c-c398854ac88d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.961716] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] c64056d8-0c22-48db-bf4a-a60f468bdad0/c64056d8-0c22-48db-bf4a-a60f468bdad0.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.962044] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-998831b2-96f0-43a9-a4a4-85f777108fce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.983809] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 881.983809] env[61985]: value = "task-935985" [ 881.983809] env[61985]: _type = "Task" [ 881.983809] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.995848] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524fd153-725c-f75e-9d67-c287289ee548, 'name': SearchDatastore_Task, 'duration_secs': 0.010709} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.999113] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.999359] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.999618] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.999793] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.999990] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.000277] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935985, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.000493] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0287a19-b394-426f-a114-c1d2006937fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.009332] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.009524] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 882.010348] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24380ffd-1956-476c-bfed-240a57be9853 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.015764] env[61985]: DEBUG nova.objects.instance [None req-ae92469e-2a71-479e-9756-9d24ad08e02f tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'pci_requests' on Instance uuid 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.018176] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 882.018176] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522636e5-bcc6-f875-7aab-612c95e59e5a" [ 882.018176] env[61985]: _type = "Task" [ 882.018176] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.027264] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522636e5-bcc6-f875-7aab-612c95e59e5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.192816] env[61985]: DEBUG oslo_vmware.api [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935982, 'name': PowerOnVM_Task, 'duration_secs': 1.023055} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.193171] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 882.193425] env[61985]: INFO nova.compute.manager [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Took 9.69 seconds to spawn the instance on the hypervisor. [ 882.193680] env[61985]: DEBUG nova.compute.manager [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.194594] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0788c09-6cb2-4058-b7b8-f2d409aa2b65 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.347816] env[61985]: DEBUG nova.compute.manager [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 882.350561] env[61985]: DEBUG nova.scheduler.client.report [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.371998] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.372325] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.372501] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.372688] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.372839] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.372991] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.373247] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.373365] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.373631] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.373702] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.373894] env[61985]: DEBUG nova.virt.hardware [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.374941] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2775eb0-ac62-4021-b6ca-5982a3780f9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.384107] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78cb055d-6bbc-441a-99f0-ad7f6342ed2d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.398729] env[61985]: DEBUG oslo_concurrency.lockutils [req-1c70178d-124c-4aca-ab7c-7bbcdcff6165 req-9b5ca5de-ce63-45d2-bb69-671ba7002547 service nova] Releasing lock "refresh_cache-e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.494220] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935985, 'name': ReconfigVM_Task, 'duration_secs': 0.329297} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.494512] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Reconfigured VM instance instance-0000003f to attach disk [datastore1] c64056d8-0c22-48db-bf4a-a60f468bdad0/c64056d8-0c22-48db-bf4a-a60f468bdad0.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.495132] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8be491dc-9c56-462e-93a1-2255cd42eea4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.502297] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 882.502297] env[61985]: value = "task-935986" [ 882.502297] env[61985]: _type = "Task" [ 882.502297] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.510580] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935986, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.519252] env[61985]: DEBUG nova.objects.base [None req-ae92469e-2a71-479e-9756-9d24ad08e02f tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Object Instance<5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf> lazy-loaded attributes: flavor,pci_requests {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 882.519464] env[61985]: DEBUG nova.network.neutron [None req-ae92469e-2a71-479e-9756-9d24ad08e02f tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 882.531235] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522636e5-bcc6-f875-7aab-612c95e59e5a, 'name': SearchDatastore_Task, 'duration_secs': 0.013406} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.531433] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3fd100f-135d-4a1a-8b68-777a0f3f9b48 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.538129] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 882.538129] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525db64c-7339-1dee-b9c3-40946e987bf1" [ 882.538129] env[61985]: _type = "Task" [ 882.538129] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.547514] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525db64c-7339-1dee-b9c3-40946e987bf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.625383] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae92469e-2a71-479e-9756-9d24ad08e02f tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.114s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.715927] env[61985]: INFO nova.compute.manager [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Took 42.33 seconds to build instance. [ 882.855564] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.856130] env[61985]: DEBUG nova.compute.manager [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 882.858963] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.128s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.862677] env[61985]: INFO nova.compute.claims [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.005195] env[61985]: DEBUG nova.compute.manager [req-cfb4c265-33f7-44e5-8bbe-35942062a349 req-1174f596-c913-40e5-b2c9-6006a7987ae2 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Received event network-changed-4d18bb14-4f82-46b7-b128-9acc3eaae43f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 883.005403] env[61985]: DEBUG nova.compute.manager [req-cfb4c265-33f7-44e5-8bbe-35942062a349 req-1174f596-c913-40e5-b2c9-6006a7987ae2 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Refreshing instance network info cache due to event network-changed-4d18bb14-4f82-46b7-b128-9acc3eaae43f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 883.005630] env[61985]: DEBUG oslo_concurrency.lockutils [req-cfb4c265-33f7-44e5-8bbe-35942062a349 req-1174f596-c913-40e5-b2c9-6006a7987ae2 service nova] Acquiring lock "refresh_cache-ce757dda-f58a-47b3-b319-e6b01f05b20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.005780] env[61985]: DEBUG oslo_concurrency.lockutils [req-cfb4c265-33f7-44e5-8bbe-35942062a349 req-1174f596-c913-40e5-b2c9-6006a7987ae2 service nova] Acquired lock "refresh_cache-ce757dda-f58a-47b3-b319-e6b01f05b20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.008158] env[61985]: DEBUG nova.network.neutron [req-cfb4c265-33f7-44e5-8bbe-35942062a349 req-1174f596-c913-40e5-b2c9-6006a7987ae2 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Refreshing network info cache for port 4d18bb14-4f82-46b7-b128-9acc3eaae43f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 883.017963] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935986, 'name': Rename_Task, 'duration_secs': 0.156708} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.020531] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 883.022389] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c07cba4-0ae1-426c-a4bf-7392c47109cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.031617] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 883.031617] env[61985]: value = "task-935987" [ 883.031617] env[61985]: _type = "Task" [ 883.031617] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.044714] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935987, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.051443] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525db64c-7339-1dee-b9c3-40946e987bf1, 'name': SearchDatastore_Task, 'duration_secs': 0.011309} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.051701] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.052068] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] e68cc5ff-83b7-4602-be50-ee4b1fa9ac65/e68cc5ff-83b7-4602-be50-ee4b1fa9ac65.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 883.052403] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c65bb636-41d0-49b2-8f3b-9fcf5d2f0082 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.059431] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 883.059431] env[61985]: value = "task-935988" [ 883.059431] env[61985]: _type = "Task" [ 883.059431] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.069671] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935988, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.219289] env[61985]: DEBUG oslo_concurrency.lockutils [None req-463908cc-fb17-48d1-860e-b5b86dd91e5a tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.178s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.311316] env[61985]: DEBUG nova.network.neutron [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Successfully updated port: cfee741d-7a1d-4e4f-8432-1603248a1ba4 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.365560] env[61985]: DEBUG nova.compute.utils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.369429] env[61985]: DEBUG nova.compute.manager [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.369631] env[61985]: DEBUG nova.network.neutron [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 883.422271] env[61985]: DEBUG nova.policy [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9ab5fa74ead24de0822ff966d103c446', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd88faeaedf145c0bea490ba1debfbd7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.544383] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935987, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.570311] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935988, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.721656] env[61985]: DEBUG nova.compute.manager [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.812885] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquiring lock "refresh_cache-518635bf-73ee-404b-ae6a-dc4ee23009d2" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.813066] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquired lock "refresh_cache-518635bf-73ee-404b-ae6a-dc4ee23009d2" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.813223] env[61985]: DEBUG nova.network.neutron [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 883.871581] env[61985]: DEBUG nova.compute.manager [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 883.993640] env[61985]: DEBUG nova.network.neutron [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Successfully created port: 8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.006034] env[61985]: DEBUG nova.network.neutron [req-cfb4c265-33f7-44e5-8bbe-35942062a349 req-1174f596-c913-40e5-b2c9-6006a7987ae2 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Updated VIF entry in instance network info cache for port 4d18bb14-4f82-46b7-b128-9acc3eaae43f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.006390] env[61985]: DEBUG nova.network.neutron [req-cfb4c265-33f7-44e5-8bbe-35942062a349 req-1174f596-c913-40e5-b2c9-6006a7987ae2 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Updating instance_info_cache with network_info: [{"id": "4d18bb14-4f82-46b7-b128-9acc3eaae43f", "address": "fa:16:3e:62:e1:57", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d18bb14-4f", "ovs_interfaceid": "4d18bb14-4f82-46b7-b128-9acc3eaae43f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.046527] env[61985]: DEBUG oslo_vmware.api [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-935987, 'name': PowerOnVM_Task, 'duration_secs': 0.944162} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.047196] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 884.047446] env[61985]: INFO nova.compute.manager [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Took 8.94 seconds to spawn the instance on the hypervisor. [ 884.047754] env[61985]: DEBUG nova.compute.manager [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.048598] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da60bd68-02b3-4cbc-96fb-884105140f11 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.070338] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935988, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556488} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.072896] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] e68cc5ff-83b7-4602-be50-ee4b1fa9ac65/e68cc5ff-83b7-4602-be50-ee4b1fa9ac65.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 884.073139] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.073571] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63c5369f-e304-4ace-9c5e-75563f43f7d5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.081707] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 884.081707] env[61985]: value = "task-935989" [ 884.081707] env[61985]: _type = "Task" [ 884.081707] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.095544] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935989, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.240236] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.307585] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215fae44-5f20-4275-8eed-feee1536d47a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.315770] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5faf8c36-ae63-4124-bbd1-dd7a4e0a4995 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.348843] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5b1141-e30b-46e2-afa4-da013bb98cb0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.357052] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7136c745-9826-48a8-a596-4fc341b87131 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.371503] env[61985]: DEBUG nova.compute.provider_tree [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.385769] env[61985]: DEBUG nova.network.neutron [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 884.508511] env[61985]: DEBUG oslo_concurrency.lockutils [req-cfb4c265-33f7-44e5-8bbe-35942062a349 req-1174f596-c913-40e5-b2c9-6006a7987ae2 service nova] Releasing lock "refresh_cache-ce757dda-f58a-47b3-b319-e6b01f05b20b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.577406] env[61985]: INFO nova.compute.manager [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Took 40.97 seconds to build instance. [ 884.591896] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935989, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.458332} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.592189] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.592991] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ac51bf-c9ec-4c8a-abb3-2b2a6c0963e0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.628794] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] e68cc5ff-83b7-4602-be50-ee4b1fa9ac65/e68cc5ff-83b7-4602-be50-ee4b1fa9ac65.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.629602] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4034490-5c98-4f7a-8dfc-2846689063f0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.645622] env[61985]: DEBUG nova.network.neutron [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Updating instance_info_cache with network_info: [{"id": "cfee741d-7a1d-4e4f-8432-1603248a1ba4", "address": "fa:16:3e:77:30:94", "network": {"id": "50a76089-e640-4de0-83e3-afb2f6cc11af", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-774539119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4ebaf2c61d940d3bab278994b54b418", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfee741d-7a", "ovs_interfaceid": "cfee741d-7a1d-4e4f-8432-1603248a1ba4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.654139] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 884.654139] env[61985]: value = "task-935990" [ 884.654139] env[61985]: _type = "Task" [ 884.654139] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.663992] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935990, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.874676] env[61985]: DEBUG nova.scheduler.client.report [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.886742] env[61985]: DEBUG nova.compute.manager [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 884.914357] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.915113] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.915367] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.915655] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.915824] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.915982] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.916216] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.916383] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.916555] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.916726] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.916906] env[61985]: DEBUG nova.virt.hardware [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.917826] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53237aa1-719f-4959-acb5-85f37af2a2da {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.926902] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8ad04c-8a0c-4214-80af-bb02df773868 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.966758] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquiring lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.966991] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.967248] env[61985]: INFO nova.compute.manager [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Rebooting instance [ 884.999904] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.000283] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.000629] env[61985]: DEBUG nova.objects.instance [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'flavor' on Instance uuid 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.079106] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9ce8878-ac21-4d8f-bf75-4dc305ae3fcb tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.510s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.147807] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Releasing lock "refresh_cache-518635bf-73ee-404b-ae6a-dc4ee23009d2" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.148169] env[61985]: DEBUG nova.compute.manager [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Instance network_info: |[{"id": "cfee741d-7a1d-4e4f-8432-1603248a1ba4", "address": "fa:16:3e:77:30:94", "network": {"id": "50a76089-e640-4de0-83e3-afb2f6cc11af", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-774539119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4ebaf2c61d940d3bab278994b54b418", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfee741d-7a", "ovs_interfaceid": "cfee741d-7a1d-4e4f-8432-1603248a1ba4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 885.148645] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:30:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cfee741d-7a1d-4e4f-8432-1603248a1ba4', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.156304] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Creating folder: Project (b4ebaf2c61d940d3bab278994b54b418). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 885.156974] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b6d0f6db-de78-4206-9f0e-6032b9c2dfb7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.169176] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935990, 'name': ReconfigVM_Task, 'duration_secs': 0.392776} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.169482] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Reconfigured VM instance instance-00000040 to attach disk [datastore1] e68cc5ff-83b7-4602-be50-ee4b1fa9ac65/e68cc5ff-83b7-4602-be50-ee4b1fa9ac65.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.170241] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af15bbba-06d7-4c91-93af-c8e3b34a8b45 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.173085] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Created folder: Project (b4ebaf2c61d940d3bab278994b54b418) in parent group-v211285. [ 885.173504] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Creating folder: Instances. Parent ref: group-v211387. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 885.174119] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46cfe825-3402-41e1-aaf1-177f313a0921 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.179523] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 885.179523] env[61985]: value = "task-935992" [ 885.179523] env[61985]: _type = "Task" [ 885.179523] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.184704] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Created folder: Instances in parent group-v211387. [ 885.184973] env[61985]: DEBUG oslo.service.loopingcall [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.185202] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 885.185469] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b45eff1a-1175-4cca-9d70-74ce78efbc9f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.205691] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935992, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.212441] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.212441] env[61985]: value = "task-935994" [ 885.212441] env[61985]: _type = "Task" [ 885.212441] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.221852] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935994, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.380273] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.380851] env[61985]: DEBUG nova.compute.manager [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 885.383714] env[61985]: DEBUG oslo_concurrency.lockutils [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.133s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.383945] env[61985]: DEBUG nova.objects.instance [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lazy-loading 'resources' on Instance uuid 95e3e6d7-9203-47a0-bad7-050eba09c511 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.482575] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquiring lock "refresh_cache-5b3a8ba3-48cb-472a-ab45-4e51af68812c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.482872] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquired lock "refresh_cache-5b3a8ba3-48cb-472a-ab45-4e51af68812c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.483093] env[61985]: DEBUG nova.network.neutron [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 885.581565] env[61985]: DEBUG nova.compute.manager [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 885.690967] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935992, 'name': Rename_Task, 'duration_secs': 0.188855} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.691352] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 885.691628] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ab3a120-bdcd-4c10-a83a-777a9e8f119d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.700680] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 885.700680] env[61985]: value = "task-935995" [ 885.700680] env[61985]: _type = "Task" [ 885.700680] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.712324] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935995, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.724440] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935994, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.754116] env[61985]: DEBUG nova.objects.instance [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'pci_requests' on Instance uuid 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.846283] env[61985]: DEBUG nova.compute.manager [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Received event network-vif-plugged-cfee741d-7a1d-4e4f-8432-1603248a1ba4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 885.846575] env[61985]: DEBUG oslo_concurrency.lockutils [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] Acquiring lock "518635bf-73ee-404b-ae6a-dc4ee23009d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.846764] env[61985]: DEBUG oslo_concurrency.lockutils [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] Lock "518635bf-73ee-404b-ae6a-dc4ee23009d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.846906] env[61985]: DEBUG oslo_concurrency.lockutils [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] Lock "518635bf-73ee-404b-ae6a-dc4ee23009d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.847100] env[61985]: DEBUG nova.compute.manager [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] No waiting events found dispatching network-vif-plugged-cfee741d-7a1d-4e4f-8432-1603248a1ba4 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 885.847278] env[61985]: WARNING nova.compute.manager [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Received unexpected event network-vif-plugged-cfee741d-7a1d-4e4f-8432-1603248a1ba4 for instance with vm_state building and task_state spawning. [ 885.847442] env[61985]: DEBUG nova.compute.manager [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Received event network-changed-cfee741d-7a1d-4e4f-8432-1603248a1ba4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 885.847632] env[61985]: DEBUG nova.compute.manager [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Refreshing instance network info cache due to event network-changed-cfee741d-7a1d-4e4f-8432-1603248a1ba4. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 885.847860] env[61985]: DEBUG oslo_concurrency.lockutils [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] Acquiring lock "refresh_cache-518635bf-73ee-404b-ae6a-dc4ee23009d2" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.848028] env[61985]: DEBUG oslo_concurrency.lockutils [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] Acquired lock "refresh_cache-518635bf-73ee-404b-ae6a-dc4ee23009d2" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.848203] env[61985]: DEBUG nova.network.neutron [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Refreshing network info cache for port cfee741d-7a1d-4e4f-8432-1603248a1ba4 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 885.886684] env[61985]: DEBUG nova.compute.utils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.888164] env[61985]: DEBUG nova.compute.manager [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.888344] env[61985]: DEBUG nova.network.neutron [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 885.948070] env[61985]: DEBUG nova.policy [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '00f4e0f073684affbdc22e676f440838', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '091a92048e3545b7a6b98af4e81d2f00', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 886.106217] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.213660] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935995, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.220441] env[61985]: DEBUG nova.network.neutron [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Updating instance_info_cache with network_info: [{"id": "e25b97ce-77e2-472d-a7a2-42899e22aab2", "address": "fa:16:3e:ea:45:2a", "network": {"id": "8d56ab44-9a85-46d5-9ce5-9bfe7830ba5f", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1009275674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "006a357baa9f4bd288cbec53a0be88cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape25b97ce-77", "ovs_interfaceid": "e25b97ce-77e2-472d-a7a2-42899e22aab2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.229207] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-935994, 'name': CreateVM_Task, 'duration_secs': 0.584189} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.230867] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 886.235121] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.235121] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.235121] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 886.235121] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8d1d96b-7547-473e-af21-099fc150c265 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.240595] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 886.240595] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52de295f-996a-e7b4-417c-02072d683807" [ 886.240595] env[61985]: _type = "Task" [ 886.240595] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.255431] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52de295f-996a-e7b4-417c-02072d683807, 'name': SearchDatastore_Task, 'duration_secs': 0.012345} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.255819] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.256107] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.256377] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.256648] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.256895] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.257492] env[61985]: DEBUG nova.objects.base [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Object Instance<5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf> lazy-loaded attributes: flavor,pci_requests {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 886.257861] env[61985]: DEBUG nova.network.neutron [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 886.259750] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cbac481a-ed65-439e-ad9d-d594bb81c0b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.274338] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.274338] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 886.275086] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3ee939c-1c70-43ea-a04e-87795f6fda00 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.288019] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 886.288019] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52123ee3-29b6-4900-156f-5fb057068dfe" [ 886.288019] env[61985]: _type = "Task" [ 886.288019] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.299615] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52123ee3-29b6-4900-156f-5fb057068dfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.319022] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b74b81-d681-4db1-bef0-d94b68f09f59 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.327971] env[61985]: DEBUG nova.policy [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a78dbc667d24c819e6447740fc2d8e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2dbceebf07c43ac90ed80f8a09294b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 886.329765] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d546469d-3734-4229-9d78-c90bb03f2fc1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.364031] env[61985]: DEBUG nova.network.neutron [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Successfully created port: 97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.366502] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24eb9c5-d296-4d3d-b440-b2f2165d2339 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.376993] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26774739-9281-488a-9f5a-0d52e4c31347 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.393835] env[61985]: DEBUG nova.compute.manager [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 886.398141] env[61985]: DEBUG nova.compute.provider_tree [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.452887] env[61985]: DEBUG nova.network.neutron [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Successfully updated port: 8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.618196] env[61985]: DEBUG nova.network.neutron [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Successfully created port: 94987e66-83ff-441e-924c-a937a0c6dc27 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.714347] env[61985]: DEBUG oslo_vmware.api [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-935995, 'name': PowerOnVM_Task, 'duration_secs': 0.728569} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.718021] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 886.718021] env[61985]: INFO nova.compute.manager [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Took 9.03 seconds to spawn the instance on the hypervisor. [ 886.718021] env[61985]: DEBUG nova.compute.manager [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.718021] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f753e470-b67e-48ff-83e6-0db2c0411b45 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.723164] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Releasing lock "refresh_cache-5b3a8ba3-48cb-472a-ab45-4e51af68812c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.727750] env[61985]: DEBUG nova.compute.manager [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.728908] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fc7bbd-cd81-4faa-841a-35c373d712ed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.798747] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52123ee3-29b6-4900-156f-5fb057068dfe, 'name': SearchDatastore_Task, 'duration_secs': 0.011475} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.799707] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-853ef548-9ab9-4c24-88e4-94806d75d0cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.808940] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 886.808940] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5216f52d-84f3-eff6-a7b5-6ece04e78bec" [ 886.808940] env[61985]: _type = "Task" [ 886.808940] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.819124] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5216f52d-84f3-eff6-a7b5-6ece04e78bec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.881032] env[61985]: DEBUG nova.network.neutron [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Updated VIF entry in instance network info cache for port cfee741d-7a1d-4e4f-8432-1603248a1ba4. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 886.881032] env[61985]: DEBUG nova.network.neutron [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Updating instance_info_cache with network_info: [{"id": "cfee741d-7a1d-4e4f-8432-1603248a1ba4", "address": "fa:16:3e:77:30:94", "network": {"id": "50a76089-e640-4de0-83e3-afb2f6cc11af", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-774539119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4ebaf2c61d940d3bab278994b54b418", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfee741d-7a", "ovs_interfaceid": "cfee741d-7a1d-4e4f-8432-1603248a1ba4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.905304] env[61985]: DEBUG nova.scheduler.client.report [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.954912] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquiring lock "refresh_cache-54ba5cfd-185a-4c58-aa5e-83cc096a482e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.955081] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquired lock "refresh_cache-54ba5cfd-185a-4c58-aa5e-83cc096a482e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.955234] env[61985]: DEBUG nova.network.neutron [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.241818] env[61985]: INFO nova.compute.manager [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Took 42.70 seconds to build instance. [ 887.321414] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5216f52d-84f3-eff6-a7b5-6ece04e78bec, 'name': SearchDatastore_Task, 'duration_secs': 0.038158} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.321843] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.322237] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 518635bf-73ee-404b-ae6a-dc4ee23009d2/518635bf-73ee-404b-ae6a-dc4ee23009d2.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 887.322664] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1d66c3a-38b7-463d-91a3-4d97b3f01be4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.330791] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 887.330791] env[61985]: value = "task-935996" [ 887.330791] env[61985]: _type = "Task" [ 887.330791] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.339908] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-935996, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.383300] env[61985]: DEBUG oslo_concurrency.lockutils [req-ea5c20fe-d7f0-4332-b76e-bcd6767dce54 req-b2a1c02f-64ca-468a-95e1-d2c1fbd5775c service nova] Releasing lock "refresh_cache-518635bf-73ee-404b-ae6a-dc4ee23009d2" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.410612] env[61985]: DEBUG nova.compute.manager [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 887.413086] env[61985]: DEBUG oslo_concurrency.lockutils [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.029s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.415915] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.119s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.415915] env[61985]: DEBUG nova.objects.instance [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lazy-loading 'resources' on Instance uuid f8924f18-7232-4be6-84bf-1ba05b855ffe {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.444472] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.444750] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.444915] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.445116] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.445270] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.445486] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.445727] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.445903] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.446077] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.446251] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.446432] env[61985]: DEBUG nova.virt.hardware [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.447673] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ff3759-71b0-4635-82f9-844748a8aba1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.451806] env[61985]: INFO nova.scheduler.client.report [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleted allocations for instance 95e3e6d7-9203-47a0-bad7-050eba09c511 [ 887.460974] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7daf3a35-f6be-429f-8495-0c94065c1f38 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.518083] env[61985]: DEBUG nova.network.neutron [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 887.743171] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e60a306-6e49-46c7-bfba-cc4cfd31a69d tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.746s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.745193] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca633af7-2110-4e58-a2c6-3d16e505cfa6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.756660] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Doing hard reboot of VM {{(pid=61985) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 887.756969] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-35dc2af8-176f-4638-b819-d2a5ecb668a8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.766156] env[61985]: DEBUG oslo_vmware.api [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 887.766156] env[61985]: value = "task-935997" [ 887.766156] env[61985]: _type = "Task" [ 887.766156] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.776436] env[61985]: DEBUG oslo_vmware.api [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935997, 'name': ResetVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.816068] env[61985]: DEBUG nova.network.neutron [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Updating instance_info_cache with network_info: [{"id": "8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4", "address": "fa:16:3e:54:65:aa", "network": {"id": "b753bbd4-bdb2-4823-afa1-93345afed5e9", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-303829824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd88faeaedf145c0bea490ba1debfbd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be34cd9-0c", "ovs_interfaceid": "8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.842167] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-935996, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.965323] env[61985]: DEBUG oslo_concurrency.lockutils [None req-feb7723f-50e2-4861-aa89-f2639efa189d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "95e3e6d7-9203-47a0-bad7-050eba09c511" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.565s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.035378] env[61985]: DEBUG nova.compute.manager [req-92c0b255-af72-4056-b143-9d607b6a1ae6 req-bd0a6263-9dd3-4f3d-9059-f762533a2b88 service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Received event network-vif-plugged-8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 888.035769] env[61985]: DEBUG oslo_concurrency.lockutils [req-92c0b255-af72-4056-b143-9d607b6a1ae6 req-bd0a6263-9dd3-4f3d-9059-f762533a2b88 service nova] Acquiring lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.036011] env[61985]: DEBUG oslo_concurrency.lockutils [req-92c0b255-af72-4056-b143-9d607b6a1ae6 req-bd0a6263-9dd3-4f3d-9059-f762533a2b88 service nova] Lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.036199] env[61985]: DEBUG oslo_concurrency.lockutils [req-92c0b255-af72-4056-b143-9d607b6a1ae6 req-bd0a6263-9dd3-4f3d-9059-f762533a2b88 service nova] Lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.036378] env[61985]: DEBUG nova.compute.manager [req-92c0b255-af72-4056-b143-9d607b6a1ae6 req-bd0a6263-9dd3-4f3d-9059-f762533a2b88 service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] No waiting events found dispatching network-vif-plugged-8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.036865] env[61985]: WARNING nova.compute.manager [req-92c0b255-af72-4056-b143-9d607b6a1ae6 req-bd0a6263-9dd3-4f3d-9059-f762533a2b88 service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Received unexpected event network-vif-plugged-8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4 for instance with vm_state building and task_state spawning. [ 888.183611] env[61985]: DEBUG nova.network.neutron [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Successfully updated port: 94987e66-83ff-441e-924c-a937a0c6dc27 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.277106] env[61985]: DEBUG oslo_vmware.api [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-935997, 'name': ResetVM_Task, 'duration_secs': 0.098907} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.280291] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Did hard reboot of VM {{(pid=61985) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 888.280291] env[61985]: DEBUG nova.compute.manager [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.281209] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e344a8-1b24-44e6-b348-412824aa95af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.322531] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Releasing lock "refresh_cache-54ba5cfd-185a-4c58-aa5e-83cc096a482e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.322849] env[61985]: DEBUG nova.compute.manager [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Instance network_info: |[{"id": "8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4", "address": "fa:16:3e:54:65:aa", "network": {"id": "b753bbd4-bdb2-4823-afa1-93345afed5e9", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-303829824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd88faeaedf145c0bea490ba1debfbd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be34cd9-0c", "ovs_interfaceid": "8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.323343] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:65:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6c45fd-e930-495a-9cb7-df84eda443b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.332587] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Creating folder: Project (fd88faeaedf145c0bea490ba1debfbd7). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 888.334701] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b5405152-664e-4903-936b-eb6056dad93e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.347234] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-935996, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543568} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.347968] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 518635bf-73ee-404b-ae6a-dc4ee23009d2/518635bf-73ee-404b-ae6a-dc4ee23009d2.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 888.348209] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 888.349344] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6d6a3cdf-b048-4afa-b3b8-958a082f60d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.351675] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049be42c-1692-4028-af3a-03ad145b1a0a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.356430] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Created folder: Project (fd88faeaedf145c0bea490ba1debfbd7) in parent group-v211285. [ 888.356621] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Creating folder: Instances. Parent ref: group-v211390. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 888.357463] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8c9717e-a23b-4b7a-a4eb-8f4abfff1c21 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.362996] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4096c96-9c7e-4f15-a5a9-cdc03f279923 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.366179] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 888.366179] env[61985]: value = "task-935999" [ 888.366179] env[61985]: _type = "Task" [ 888.366179] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.367505] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Created folder: Instances in parent group-v211390. [ 888.367777] env[61985]: DEBUG oslo.service.loopingcall [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.371533] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 888.399813] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de95bbb5-0a5f-4f82-9130-8210f9175a2e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.416306] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c9dc8f-ab4e-47dc-999f-2749aa9fd6d7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.424018] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-935999, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.427781] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11430524-af73-4436-89a4-bb1a69e21dc8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.434209] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.434209] env[61985]: value = "task-936001" [ 888.434209] env[61985]: _type = "Task" [ 888.434209] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.437094] env[61985]: DEBUG nova.network.neutron [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Successfully updated port: 97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.446122] env[61985]: DEBUG nova.compute.provider_tree [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.455689] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936001, 'name': CreateVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.696318] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.696524] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.696759] env[61985]: DEBUG nova.network.neutron [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 888.802449] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27a28339-5dfc-44da-9c84-e96b3e9851c5 tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.835s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.877730] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-935999, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073129} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.878042] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.878859] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb91d185-2d7e-4b2c-b779-86952e565468 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.901458] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 518635bf-73ee-404b-ae6a-dc4ee23009d2/518635bf-73ee-404b-ae6a-dc4ee23009d2.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.901744] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bea1960-433b-4255-87fc-7c880c8e441f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.923160] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 888.923160] env[61985]: value = "task-936002" [ 888.923160] env[61985]: _type = "Task" [ 888.923160] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.931852] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936002, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.943488] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936001, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.948736] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.948871] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.949039] env[61985]: DEBUG nova.network.neutron [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 888.950637] env[61985]: DEBUG nova.scheduler.client.report [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.194930] env[61985]: DEBUG nova.compute.manager [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Received event network-changed-8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 889.195153] env[61985]: DEBUG nova.compute.manager [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Refreshing instance network info cache due to event network-changed-8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 889.195386] env[61985]: DEBUG oslo_concurrency.lockutils [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] Acquiring lock "refresh_cache-54ba5cfd-185a-4c58-aa5e-83cc096a482e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.195538] env[61985]: DEBUG oslo_concurrency.lockutils [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] Acquired lock "refresh_cache-54ba5cfd-185a-4c58-aa5e-83cc096a482e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.195707] env[61985]: DEBUG nova.network.neutron [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Refreshing network info cache for port 8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 889.270529] env[61985]: WARNING nova.network.neutron [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] 5014dbd4-285f-43e0-ad95-2107c34e808a already exists in list: networks containing: ['5014dbd4-285f-43e0-ad95-2107c34e808a']. ignoring it [ 889.434905] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936002, 'name': ReconfigVM_Task, 'duration_secs': 0.480883} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.438122] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 518635bf-73ee-404b-ae6a-dc4ee23009d2/518635bf-73ee-404b-ae6a-dc4ee23009d2.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.438766] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-484b65f4-2fc9-4fa2-8328-a041fe30f43e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.446114] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936001, 'name': CreateVM_Task, 'duration_secs': 0.728747} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.447372] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 889.447750] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 889.447750] env[61985]: value = "task-936003" [ 889.447750] env[61985]: _type = "Task" [ 889.447750] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.448404] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.448569] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.448908] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 889.449236] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-397d362c-54ef-4712-878a-101f819e4bb4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.456528] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.041s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.462787] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.469s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.463059] env[61985]: DEBUG nova.objects.instance [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lazy-loading 'resources' on Instance uuid e82d26b1-5502-4fd2-89c5-ffb0c1557c79 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.465033] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for the task: (returnval){ [ 889.465033] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c99877-6a29-c9e3-1268-7699702f0eae" [ 889.465033] env[61985]: _type = "Task" [ 889.465033] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.465033] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936003, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.478801] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c99877-6a29-c9e3-1268-7699702f0eae, 'name': SearchDatastore_Task, 'duration_secs': 0.011071} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.479817] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.480136] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 889.480386] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.480646] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.480936] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.481509] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2fda4a52-ec90-46f3-bf2c-28ef6bfb79d9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.488188] env[61985]: INFO nova.scheduler.client.report [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Deleted allocations for instance f8924f18-7232-4be6-84bf-1ba05b855ffe [ 889.494686] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.494872] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 889.497174] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaea9f14-39cd-4532-887c-a756997d603d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.505280] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for the task: (returnval){ [ 889.505280] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527ae4d8-db5e-e35e-eac3-18df2a8554be" [ 889.505280] env[61985]: _type = "Task" [ 889.505280] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.509615] env[61985]: DEBUG nova.network.neutron [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 889.517881] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527ae4d8-db5e-e35e-eac3-18df2a8554be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.657311] env[61985]: DEBUG nova.network.neutron [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updating instance_info_cache with network_info: [{"id": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "address": "fa:16:3e:fc:45:f4", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97f641fb-39", "ovs_interfaceid": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.791013] env[61985]: DEBUG nova.network.neutron [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updating instance_info_cache with network_info: [{"id": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "address": "fa:16:3e:68:94:dd", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4959d778-77", "ovs_interfaceid": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "94987e66-83ff-441e-924c-a937a0c6dc27", "address": "fa:16:3e:40:92:00", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94987e66-83", "ovs_interfaceid": "94987e66-83ff-441e-924c-a937a0c6dc27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.893152] env[61985]: DEBUG nova.network.neutron [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Updated VIF entry in instance network info cache for port 8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 889.893534] env[61985]: DEBUG nova.network.neutron [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Updating instance_info_cache with network_info: [{"id": "8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4", "address": "fa:16:3e:54:65:aa", "network": {"id": "b753bbd4-bdb2-4823-afa1-93345afed5e9", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-303829824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd88faeaedf145c0bea490ba1debfbd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be34cd9-0c", "ovs_interfaceid": "8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.959761] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936003, 'name': Rename_Task, 'duration_secs': 0.15943} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.960064] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 889.960419] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3455e797-bca3-4459-8a8a-f61893a17e04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.970873] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 889.970873] env[61985]: value = "task-936004" [ 889.970873] env[61985]: _type = "Task" [ 889.970873] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.979645] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936004, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.000423] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9cdb022b-87b9-48c9-b07d-ae0714a0769b tempest-ServersAaction247Test-1298778519 tempest-ServersAaction247Test-1298778519-project-member] Lock "f8924f18-7232-4be6-84bf-1ba05b855ffe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.440s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.025770] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527ae4d8-db5e-e35e-eac3-18df2a8554be, 'name': SearchDatastore_Task, 'duration_secs': 0.011355} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.030404] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4d4ede3-3ed2-4b3a-b79a-60ff2e237ad6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.037661] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for the task: (returnval){ [ 890.037661] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520713cb-78d4-3d36-10b6-f898b5bb2a00" [ 890.037661] env[61985]: _type = "Task" [ 890.037661] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.053012] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520713cb-78d4-3d36-10b6-f898b5bb2a00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.164553] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.164970] env[61985]: DEBUG nova.compute.manager [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Instance network_info: |[{"id": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "address": "fa:16:3e:fc:45:f4", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97f641fb-39", "ovs_interfaceid": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 890.165546] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:45:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cd38fdec-d092-4a84-ab41-685f6dbb4f29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97f641fb-39ac-441c-bacc-6705d1ea6e98', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.175127] env[61985]: DEBUG oslo.service.loopingcall [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.178075] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 890.178495] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-76b1ab6a-5fb3-4c70-871d-56db06cf9aa4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.202172] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.202172] env[61985]: value = "task-936005" [ 890.202172] env[61985]: _type = "Task" [ 890.202172] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.213330] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936005, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.293348] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.293970] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.294166] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.295014] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c3bb28-fb33-4cd1-848a-a067b8898f0f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.313974] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 890.315211] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 890.315211] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 890.315211] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 890.315211] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 890.315211] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 890.315211] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 890.315562] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 890.315562] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 890.315651] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 890.315798] env[61985]: DEBUG nova.virt.hardware [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 890.322057] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Reconfiguring VM to attach interface {{(pid=61985) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 890.324917] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75b6706f-08ee-4158-9a4b-acc5552540e2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.343740] env[61985]: DEBUG oslo_vmware.api [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 890.343740] env[61985]: value = "task-936006" [ 890.343740] env[61985]: _type = "Task" [ 890.343740] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.355299] env[61985]: DEBUG oslo_vmware.api [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936006, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.376821] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1161ff16-9297-46c6-911b-506d5aa662d9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.385876] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4486a986-5074-4fc8-ba9c-717c93c15c23 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.422940] env[61985]: DEBUG oslo_concurrency.lockutils [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] Releasing lock "refresh_cache-54ba5cfd-185a-4c58-aa5e-83cc096a482e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.423247] env[61985]: DEBUG nova.compute.manager [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Received event network-changed-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 890.423423] env[61985]: DEBUG nova.compute.manager [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Refreshing instance network info cache due to event network-changed-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 890.423641] env[61985]: DEBUG oslo_concurrency.lockutils [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] Acquiring lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.423792] env[61985]: DEBUG oslo_concurrency.lockutils [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] Acquired lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.423959] env[61985]: DEBUG nova.network.neutron [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Refreshing network info cache for port 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 890.429656] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9fd41e5-4467-495a-8154-1dc687b1f83c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.436278] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b5b6b6-75dc-4cdd-96d7-de3b00f5cc97 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.454660] env[61985]: DEBUG nova.compute.provider_tree [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.481655] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936004, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.551027] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520713cb-78d4-3d36-10b6-f898b5bb2a00, 'name': SearchDatastore_Task, 'duration_secs': 0.034186} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.551027] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.551027] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 54ba5cfd-185a-4c58-aa5e-83cc096a482e/54ba5cfd-185a-4c58-aa5e-83cc096a482e.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 890.551027] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d2ba047-88bc-4670-bf0b-dcf5251d65e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.558424] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for the task: (returnval){ [ 890.558424] env[61985]: value = "task-936007" [ 890.558424] env[61985]: _type = "Task" [ 890.558424] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.567876] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936007, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.717537] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936005, 'name': CreateVM_Task, 'duration_secs': 0.40445} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.717537] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 890.719082] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.719082] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.719082] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.719082] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e6b796c-bf5b-46e8-9756-bc0fe92601e8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.724905] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 890.724905] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525d181d-4805-30c8-83db-34f92af5c1c0" [ 890.724905] env[61985]: _type = "Task" [ 890.724905] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.738656] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525d181d-4805-30c8-83db-34f92af5c1c0, 'name': SearchDatastore_Task, 'duration_secs': 0.009627} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.739060] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.739438] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.739813] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.740118] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.740444] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.740799] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac932dd4-b278-4b9a-ac71-90112cb5f52d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.749337] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.749521] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 890.750297] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f23e7bf-105f-425e-9e20-d5eb77b9a3b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.756361] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 890.756361] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523c81d2-47fa-58c9-5b4c-91ad8a377d25" [ 890.756361] env[61985]: _type = "Task" [ 890.756361] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.767346] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523c81d2-47fa-58c9-5b4c-91ad8a377d25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.856513] env[61985]: DEBUG oslo_vmware.api [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936006, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.960027] env[61985]: DEBUG nova.scheduler.client.report [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.983980] env[61985]: DEBUG oslo_vmware.api [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936004, 'name': PowerOnVM_Task, 'duration_secs': 0.720756} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.984992] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 890.985230] env[61985]: INFO nova.compute.manager [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Took 8.64 seconds to spawn the instance on the hypervisor. [ 890.985420] env[61985]: DEBUG nova.compute.manager [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.986350] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02274e10-94bb-4457-af58-74aa998bdd87 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.041268] env[61985]: DEBUG nova.compute.manager [req-92745380-8284-4b09-8025-1a6b5187651d req-ab51c120-17f8-4781-a613-35b18f8301be service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received event network-vif-plugged-94987e66-83ff-441e-924c-a937a0c6dc27 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 891.041268] env[61985]: DEBUG oslo_concurrency.lockutils [req-92745380-8284-4b09-8025-1a6b5187651d req-ab51c120-17f8-4781-a613-35b18f8301be service nova] Acquiring lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.041372] env[61985]: DEBUG oslo_concurrency.lockutils [req-92745380-8284-4b09-8025-1a6b5187651d req-ab51c120-17f8-4781-a613-35b18f8301be service nova] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.041515] env[61985]: DEBUG oslo_concurrency.lockutils [req-92745380-8284-4b09-8025-1a6b5187651d req-ab51c120-17f8-4781-a613-35b18f8301be service nova] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.041690] env[61985]: DEBUG nova.compute.manager [req-92745380-8284-4b09-8025-1a6b5187651d req-ab51c120-17f8-4781-a613-35b18f8301be service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] No waiting events found dispatching network-vif-plugged-94987e66-83ff-441e-924c-a937a0c6dc27 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 891.041860] env[61985]: WARNING nova.compute.manager [req-92745380-8284-4b09-8025-1a6b5187651d req-ab51c120-17f8-4781-a613-35b18f8301be service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received unexpected event network-vif-plugged-94987e66-83ff-441e-924c-a937a0c6dc27 for instance with vm_state active and task_state None. [ 891.071402] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936007, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.154692] env[61985]: DEBUG nova.network.neutron [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updated VIF entry in instance network info cache for port 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 891.154955] env[61985]: DEBUG nova.network.neutron [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updating instance_info_cache with network_info: [{"id": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "address": "fa:16:3e:d4:ac:00", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60c2bc37-31", "ovs_interfaceid": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.269301] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523c81d2-47fa-58c9-5b4c-91ad8a377d25, 'name': SearchDatastore_Task, 'duration_secs': 0.009921} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.270126] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6b93557-038f-4fd3-b66d-26ca24bce91b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.277199] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 891.277199] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e9c4f0-6884-62f4-0e99-25aaa71040ef" [ 891.277199] env[61985]: _type = "Task" [ 891.277199] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.285995] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e9c4f0-6884-62f4-0e99-25aaa71040ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.356020] env[61985]: DEBUG oslo_vmware.api [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936006, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.465162] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.002s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.467748] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.372s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.468014] env[61985]: DEBUG nova.objects.instance [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lazy-loading 'resources' on Instance uuid 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.492399] env[61985]: INFO nova.scheduler.client.report [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleted allocations for instance e82d26b1-5502-4fd2-89c5-ffb0c1557c79 [ 891.509030] env[61985]: INFO nova.compute.manager [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Took 41.49 seconds to build instance. [ 891.572171] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936007, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587873} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.572527] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 54ba5cfd-185a-4c58-aa5e-83cc096a482e/54ba5cfd-185a-4c58-aa5e-83cc096a482e.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 891.572817] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 891.573474] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e82d6734-1bf7-4594-8d3d-ab3d393f8c1e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.582950] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for the task: (returnval){ [ 891.582950] env[61985]: value = "task-936008" [ 891.582950] env[61985]: _type = "Task" [ 891.582950] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.593988] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936008, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.657856] env[61985]: DEBUG oslo_concurrency.lockutils [req-4e13dbb3-81a7-4487-925d-7cafafa885a5 req-d7d417a2-c044-4e77-81aa-b575a835bfda service nova] Releasing lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.791782] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e9c4f0-6884-62f4-0e99-25aaa71040ef, 'name': SearchDatastore_Task, 'duration_secs': 0.02682} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.791974] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.792258] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 6dd8218b-c14d-40c9-87df-097fab06c669/6dd8218b-c14d-40c9-87df-097fab06c669.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 891.792571] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1e4df95-6aec-4c32-aefc-c3ba409420ad {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.801507] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 891.801507] env[61985]: value = "task-936009" [ 891.801507] env[61985]: _type = "Task" [ 891.801507] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.814416] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936009, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.856067] env[61985]: DEBUG oslo_vmware.api [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936006, 'name': ReconfigVM_Task, 'duration_secs': 1.385949} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.856409] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.856658] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Reconfigured VM to attach interface {{(pid=61985) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 892.000334] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be9ed924-4fcd-44c5-b638-cd7a74cb7af9 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "e82d26b1-5502-4fd2-89c5-ffb0c1557c79" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.179s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.011958] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3db09fef-ab8b-4778-ad39-c5a30cd74f11 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lock "518635bf-73ee-404b-ae6a-dc4ee23009d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.475s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.095669] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936008, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081909} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.096036] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.096971] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd367562-c9c1-4740-9768-bffc73a47119 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.119896] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 54ba5cfd-185a-4c58-aa5e-83cc096a482e/54ba5cfd-185a-4c58-aa5e-83cc096a482e.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.122760] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ef0a213-cc34-42fe-b192-354ddea97bbd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.152946] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for the task: (returnval){ [ 892.152946] env[61985]: value = "task-936010" [ 892.152946] env[61985]: _type = "Task" [ 892.152946] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.169647] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936010, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.312857] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936009, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.337656] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aec0196-843b-4bb7-9264-af4398a5745c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.350581] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb661c9-76c3-4501-9827-eda5d6dbca76 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.382998] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b34c63aa-615b-4744-ae3b-c83b169bfceb tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.383s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.384921] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11973cdf-dbe3-43a8-b7ca-fac6b69e71b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.394030] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f927969-ba8b-4abe-81ad-c88b9ab14e52 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.410189] env[61985]: DEBUG nova.compute.provider_tree [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.671434] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936010, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.812469] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936009, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.943708} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.812773] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 6dd8218b-c14d-40c9-87df-097fab06c669/6dd8218b-c14d-40c9-87df-097fab06c669.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 892.813018] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.813297] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd6f4cb4-effc-4e26-8890-352ffa5364da {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.822035] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 892.822035] env[61985]: value = "task-936011" [ 892.822035] env[61985]: _type = "Task" [ 892.822035] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.830603] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936011, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.912778] env[61985]: DEBUG nova.scheduler.client.report [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.090489] env[61985]: DEBUG nova.compute.manager [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Received event network-vif-plugged-97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 893.090743] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Acquiring lock "6dd8218b-c14d-40c9-87df-097fab06c669-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.090961] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Lock "6dd8218b-c14d-40c9-87df-097fab06c669-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.091158] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Lock "6dd8218b-c14d-40c9-87df-097fab06c669-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.091336] env[61985]: DEBUG nova.compute.manager [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] No waiting events found dispatching network-vif-plugged-97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 893.091590] env[61985]: WARNING nova.compute.manager [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Received unexpected event network-vif-plugged-97f641fb-39ac-441c-bacc-6705d1ea6e98 for instance with vm_state building and task_state spawning. [ 893.091708] env[61985]: DEBUG nova.compute.manager [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Received event network-changed-97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 893.091832] env[61985]: DEBUG nova.compute.manager [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Refreshing instance network info cache due to event network-changed-97f641fb-39ac-441c-bacc-6705d1ea6e98. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 893.092085] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Acquiring lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.092235] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Acquired lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.092406] env[61985]: DEBUG nova.network.neutron [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Refreshing network info cache for port 97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 893.165996] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936010, 'name': ReconfigVM_Task, 'duration_secs': 0.945608} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.167482] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 54ba5cfd-185a-4c58-aa5e-83cc096a482e/54ba5cfd-185a-4c58-aa5e-83cc096a482e.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.167482] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-118a1b1a-50f5-42d8-a1e8-84c37bf24f04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.176377] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for the task: (returnval){ [ 893.176377] env[61985]: value = "task-936012" [ 893.176377] env[61985]: _type = "Task" [ 893.176377] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.187339] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936012, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.335256] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936011, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.229056} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.335256] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.335256] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5387195-f2c6-4e47-a728-46b59b6e5ae4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.362446] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 6dd8218b-c14d-40c9-87df-097fab06c669/6dd8218b-c14d-40c9-87df-097fab06c669.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.363324] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-858d2643-45d3-4257-aa7b-4be8d2542c73 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.385054] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 893.385054] env[61985]: value = "task-936013" [ 893.385054] env[61985]: _type = "Task" [ 893.385054] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.395874] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936013, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.419031] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.951s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.421468] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.567s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.421770] env[61985]: DEBUG nova.objects.instance [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lazy-loading 'resources' on Instance uuid c4e95f51-9acc-46e8-9921-e0a85fb38598 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.447479] env[61985]: INFO nova.scheduler.client.report [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Deleted allocations for instance 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475 [ 893.689383] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936012, 'name': Rename_Task, 'duration_secs': 0.18804} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.689723] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 893.690017] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54fb7ec3-6f2b-4037-b2c5-249347ce547b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.699291] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for the task: (returnval){ [ 893.699291] env[61985]: value = "task-936014" [ 893.699291] env[61985]: _type = "Task" [ 893.699291] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.712077] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936014, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.725656] env[61985]: DEBUG nova.compute.manager [req-3043f9b6-2ecb-490b-b254-d72bc8a5d4b4 req-3dccde45-3c24-446c-a6ff-f006a398349c service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received event network-changed-94987e66-83ff-441e-924c-a937a0c6dc27 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 893.725656] env[61985]: DEBUG nova.compute.manager [req-3043f9b6-2ecb-490b-b254-d72bc8a5d4b4 req-3dccde45-3c24-446c-a6ff-f006a398349c service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Refreshing instance network info cache due to event network-changed-94987e66-83ff-441e-924c-a937a0c6dc27. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 893.725656] env[61985]: DEBUG oslo_concurrency.lockutils [req-3043f9b6-2ecb-490b-b254-d72bc8a5d4b4 req-3dccde45-3c24-446c-a6ff-f006a398349c service nova] Acquiring lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.725656] env[61985]: DEBUG oslo_concurrency.lockutils [req-3043f9b6-2ecb-490b-b254-d72bc8a5d4b4 req-3dccde45-3c24-446c-a6ff-f006a398349c service nova] Acquired lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.725656] env[61985]: DEBUG nova.network.neutron [req-3043f9b6-2ecb-490b-b254-d72bc8a5d4b4 req-3dccde45-3c24-446c-a6ff-f006a398349c service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Refreshing network info cache for port 94987e66-83ff-441e-924c-a937a0c6dc27 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 893.821526] env[61985]: DEBUG nova.network.neutron [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updated VIF entry in instance network info cache for port 97f641fb-39ac-441c-bacc-6705d1ea6e98. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 893.822134] env[61985]: DEBUG nova.network.neutron [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updating instance_info_cache with network_info: [{"id": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "address": "fa:16:3e:fc:45:f4", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97f641fb-39", "ovs_interfaceid": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.897195] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936013, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.960071] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1392e674-d9c9-4375-a538-b0af78103574 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "83e4ecc0-cc96-48e1-8c1d-7dcf84d72475" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.771s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.068527] env[61985]: DEBUG nova.compute.manager [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.069984] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16aa4dc2-7854-4c38-b7df-5f3dceeff658 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.212457] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936014, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.326976] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Releasing lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.326976] env[61985]: DEBUG nova.compute.manager [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Received event network-changed-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 894.326976] env[61985]: DEBUG nova.compute.manager [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Refreshing instance network info cache due to event network-changed-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 894.326976] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Acquiring lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.326976] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Acquired lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.326976] env[61985]: DEBUG nova.network.neutron [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Refreshing network info cache for port 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 894.370670] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34ba3cc-08c1-4d31-93c0-48366a28ec8d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.381621] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5100d1f0-9ca1-47fb-b4d2-8da3e0a9b25d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.426477] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ccd1d37-1ca1-4849-aad6-25be6642762d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.429709] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936013, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.436677] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4bffb6-954f-40e3-a377-5eb4eec12828 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.454238] env[61985]: DEBUG nova.compute.provider_tree [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.529503] env[61985]: DEBUG nova.network.neutron [req-3043f9b6-2ecb-490b-b254-d72bc8a5d4b4 req-3dccde45-3c24-446c-a6ff-f006a398349c service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updated VIF entry in instance network info cache for port 94987e66-83ff-441e-924c-a937a0c6dc27. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 894.530107] env[61985]: DEBUG nova.network.neutron [req-3043f9b6-2ecb-490b-b254-d72bc8a5d4b4 req-3dccde45-3c24-446c-a6ff-f006a398349c service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updating instance_info_cache with network_info: [{"id": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "address": "fa:16:3e:68:94:dd", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4959d778-77", "ovs_interfaceid": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "94987e66-83ff-441e-924c-a937a0c6dc27", "address": "fa:16:3e:40:92:00", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94987e66-83", "ovs_interfaceid": "94987e66-83ff-441e-924c-a937a0c6dc27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.586228] env[61985]: INFO nova.compute.manager [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] instance snapshotting [ 894.589279] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fefad82-f33d-4d5c-bec9-9224d39c859a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.609473] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3832463-f443-4705-a908-56f914fed1f4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.714063] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936014, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.788528] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquiring lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.788868] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.789609] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquiring lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.789970] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.790068] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.792790] env[61985]: INFO nova.compute.manager [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Terminating instance [ 894.794932] env[61985]: DEBUG nova.compute.manager [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 894.795164] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 894.795997] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3102910e-9fc8-44eb-9eb6-3cbd8d9d1508 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.805489] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 894.805638] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c3577633-5bb5-4153-90e5-462bc2c4697c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.814293] env[61985]: DEBUG oslo_vmware.api [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 894.814293] env[61985]: value = "task-936015" [ 894.814293] env[61985]: _type = "Task" [ 894.814293] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.826186] env[61985]: DEBUG oslo_vmware.api [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-936015, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.869959] env[61985]: DEBUG oslo_concurrency.lockutils [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "2322ab72-9841-41fb-9d60-2812baabe108" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.870339] env[61985]: DEBUG oslo_concurrency.lockutils [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.870539] env[61985]: DEBUG nova.compute.manager [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.871568] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8414f097-0619-4563-9e58-0d80c25875d9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.878789] env[61985]: DEBUG nova.compute.manager [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61985) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3374}} [ 894.879390] env[61985]: DEBUG nova.objects.instance [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'flavor' on Instance uuid 2322ab72-9841-41fb-9d60-2812baabe108 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.896485] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936013, 'name': ReconfigVM_Task, 'duration_secs': 1.318195} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.897482] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 6dd8218b-c14d-40c9-87df-097fab06c669/6dd8218b-c14d-40c9-87df-097fab06c669.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.898179] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-085cce75-4e21-4cab-a0dd-c56740fbe6ac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.905821] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 894.905821] env[61985]: value = "task-936016" [ 894.905821] env[61985]: _type = "Task" [ 894.905821] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.917708] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936016, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.955887] env[61985]: DEBUG nova.scheduler.client.report [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.033440] env[61985]: DEBUG oslo_concurrency.lockutils [req-3043f9b6-2ecb-490b-b254-d72bc8a5d4b4 req-3dccde45-3c24-446c-a6ff-f006a398349c service nova] Releasing lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.122342] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 895.122691] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3590f1dc-3baa-43b9-b377-7271927cb3d7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.133061] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 895.133061] env[61985]: value = "task-936017" [ 895.133061] env[61985]: _type = "Task" [ 895.133061] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.148283] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936017, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.157298] env[61985]: DEBUG nova.network.neutron [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updated VIF entry in instance network info cache for port 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 895.157769] env[61985]: DEBUG nova.network.neutron [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updating instance_info_cache with network_info: [{"id": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "address": "fa:16:3e:d4:ac:00", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60c2bc37-31", "ovs_interfaceid": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.212938] env[61985]: DEBUG oslo_vmware.api [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936014, 'name': PowerOnVM_Task, 'duration_secs': 1.131546} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.213279] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 895.213493] env[61985]: INFO nova.compute.manager [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Took 10.33 seconds to spawn the instance on the hypervisor. [ 895.213715] env[61985]: DEBUG nova.compute.manager [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.214652] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb3dcb0-cad3-4a9f-8282-4e7c50b0dc2a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.330242] env[61985]: DEBUG oslo_vmware.api [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-936015, 'name': PowerOffVM_Task, 'duration_secs': 0.313739} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.331259] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 895.331508] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 895.331935] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e93a52c5-89af-4d8c-8043-d7e2b50c400b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.385048] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 895.385346] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aeb1996f-672d-4450-9056-5ec673946dc7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.394281] env[61985]: DEBUG oslo_vmware.api [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 895.394281] env[61985]: value = "task-936019" [ 895.394281] env[61985]: _type = "Task" [ 895.394281] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.403848] env[61985]: DEBUG oslo_vmware.api [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936019, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.416665] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936016, 'name': Rename_Task, 'duration_secs': 0.185952} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.417435] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 895.417921] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ccbf31be-1b4c-4c96-abe9-e162092f420d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.427631] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 895.427631] env[61985]: value = "task-936020" [ 895.427631] env[61985]: _type = "Task" [ 895.427631] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.447325] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936020, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.457261] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 895.457261] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 895.457261] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Deleting the datastore file [datastore1] 5b3a8ba3-48cb-472a-ab45-4e51af68812c {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.457261] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4dbde05-5d98-4e23-b4df-ad7c619e7b7a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.465186] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.044s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.471021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.538s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.471484] env[61985]: DEBUG nova.objects.instance [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lazy-loading 'resources' on Instance uuid 4a94a6f9-1c86-4628-aa63-341f2c114e2a {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.473481] env[61985]: DEBUG oslo_vmware.api [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for the task: (returnval){ [ 895.473481] env[61985]: value = "task-936021" [ 895.473481] env[61985]: _type = "Task" [ 895.473481] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.485770] env[61985]: DEBUG oslo_vmware.api [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-936021, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.498660] env[61985]: INFO nova.scheduler.client.report [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleted allocations for instance c4e95f51-9acc-46e8-9921-e0a85fb38598 [ 895.567629] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "62824eac-9412-466a-abcf-1010f6a829e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.567830] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "62824eac-9412-466a-abcf-1010f6a829e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.618391] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.618899] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.648049] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936017, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.664587] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Releasing lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.664965] env[61985]: DEBUG nova.compute.manager [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Received event network-changed-885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 895.665243] env[61985]: DEBUG nova.compute.manager [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Refreshing instance network info cache due to event network-changed-885cda89-b90d-499a-a3b1-7155c07a1ce7. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 895.665547] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Acquiring lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.665778] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Acquired lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.665974] env[61985]: DEBUG nova.network.neutron [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Refreshing network info cache for port 885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 895.670651] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "c64056d8-0c22-48db-bf4a-a60f468bdad0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.670651] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.670651] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "c64056d8-0c22-48db-bf4a-a60f468bdad0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.670651] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.670651] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.672041] env[61985]: INFO nova.compute.manager [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Terminating instance [ 895.675539] env[61985]: DEBUG nova.compute.manager [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 895.675853] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 895.677980] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb1d10e5-206b-4f63-ba48-aa870d9d5646 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.688324] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 895.689581] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79d3a4c1-7668-4a01-aacf-e60706a4796a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.699567] env[61985]: DEBUG oslo_vmware.api [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 895.699567] env[61985]: value = "task-936022" [ 895.699567] env[61985]: _type = "Task" [ 895.699567] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.710458] env[61985]: DEBUG oslo_vmware.api [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-936022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.743773] env[61985]: INFO nova.compute.manager [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Took 43.12 seconds to build instance. [ 895.842251] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-b464350a-1222-4ab5-9caf-df714dc8f763" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.842796] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-b464350a-1222-4ab5-9caf-df714dc8f763" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.843296] env[61985]: DEBUG nova.objects.instance [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'flavor' on Instance uuid 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.908122] env[61985]: DEBUG oslo_vmware.api [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936019, 'name': PowerOffVM_Task, 'duration_secs': 0.247543} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.908683] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 895.908953] env[61985]: DEBUG nova.compute.manager [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.909909] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274846a8-ef75-4438-958c-1bdd3f7f3989 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.939414] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936020, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.985135] env[61985]: DEBUG nova.compute.manager [req-5738acbc-f293-4997-91ea-9d75576bc756 req-c2a0e4ed-f77b-4923-b586-de4a82eccf1f service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Received event network-changed-885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 895.985405] env[61985]: DEBUG nova.compute.manager [req-5738acbc-f293-4997-91ea-9d75576bc756 req-c2a0e4ed-f77b-4923-b586-de4a82eccf1f service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Refreshing instance network info cache due to event network-changed-885cda89-b90d-499a-a3b1-7155c07a1ce7. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 895.985643] env[61985]: DEBUG oslo_concurrency.lockutils [req-5738acbc-f293-4997-91ea-9d75576bc756 req-c2a0e4ed-f77b-4923-b586-de4a82eccf1f service nova] Acquiring lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.995152] env[61985]: DEBUG oslo_vmware.api [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Task: {'id': task-936021, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29965} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.995521] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.995716] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 895.995979] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 895.996149] env[61985]: INFO nova.compute.manager [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Took 1.20 seconds to destroy the instance on the hypervisor. [ 895.996417] env[61985]: DEBUG oslo.service.loopingcall [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.996631] env[61985]: DEBUG nova.compute.manager [-] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.996791] env[61985]: DEBUG nova.network.neutron [-] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 896.008049] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4700cc8d-a470-434b-aaf4-3b35b053189b tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "c4e95f51-9acc-46e8-9921-e0a85fb38598" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.065s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.071334] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 896.121845] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 896.147653] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936017, 'name': CreateSnapshot_Task, 'duration_secs': 0.971893} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.150474] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 896.151659] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5954c68c-a9fb-4d58-8268-7316258abdda {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.210523] env[61985]: DEBUG oslo_vmware.api [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-936022, 'name': PowerOffVM_Task, 'duration_secs': 0.277847} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.213260] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 896.213448] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 896.214209] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6d9ba1e-fff2-4a99-82e9-a60718ee09be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.246123] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f4bc6b48-8906-4feb-9bdf-6062f6d6944a tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.675s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.312107] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 896.312291] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 896.312468] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Deleting the datastore file [datastore1] c64056d8-0c22-48db-bf4a-a60f468bdad0 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 896.312768] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84fe14b0-a708-42b9-b983-b1cab9aa3069 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.326146] env[61985]: DEBUG oslo_vmware.api [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 896.326146] env[61985]: value = "task-936024" [ 896.326146] env[61985]: _type = "Task" [ 896.326146] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.339976] env[61985]: DEBUG oslo_vmware.api [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-936024, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.416250] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14651d7f-aa96-4b56-b8f1-fbd0c3259a67 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.425216] env[61985]: DEBUG oslo_concurrency.lockutils [None req-323cbae8-47cc-4d0f-831e-4b60139a6797 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.555s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.433669] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f4e5fe-7976-44cd-8c11-af5166f00f5e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.442406] env[61985]: DEBUG oslo_vmware.api [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936020, 'name': PowerOnVM_Task, 'duration_secs': 0.694533} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.468922] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.469238] env[61985]: INFO nova.compute.manager [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Took 9.06 seconds to spawn the instance on the hypervisor. [ 896.469454] env[61985]: DEBUG nova.compute.manager [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.470856] env[61985]: INFO nova.compute.manager [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Rebuilding instance [ 896.473899] env[61985]: DEBUG nova.objects.instance [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'pci_requests' on Instance uuid 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.477718] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8db974-141d-4a4a-8d83-c2863c5d301d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.481278] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa578db1-e718-42e3-b1f9-15bf0ea204f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.499455] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6087917-7f78-4419-b702-8c2916e7a754 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.520251] env[61985]: DEBUG nova.compute.provider_tree [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 896.525383] env[61985]: DEBUG nova.compute.manager [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.526039] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e47bb2e-6a19-4aba-b7ea-fd89aa564c95 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.593918] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.623171] env[61985]: DEBUG nova.network.neutron [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Updated VIF entry in instance network info cache for port 885cda89-b90d-499a-a3b1-7155c07a1ce7. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 896.623563] env[61985]: DEBUG nova.network.neutron [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Updating instance_info_cache with network_info: [{"id": "885cda89-b90d-499a-a3b1-7155c07a1ce7", "address": "fa:16:3e:7d:a3:a3", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap885cda89-b9", "ovs_interfaceid": "885cda89-b90d-499a-a3b1-7155c07a1ce7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.643362] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.674298] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 896.675050] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-018de157-a672-44df-9a60-25569d81fdc4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.686749] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 896.686749] env[61985]: value = "task-936025" [ 896.686749] env[61985]: _type = "Task" [ 896.686749] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.697280] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936025, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.804554] env[61985]: DEBUG nova.network.neutron [-] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.838583] env[61985]: DEBUG oslo_vmware.api [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-936024, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.463981} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.839061] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.839197] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 896.839389] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 896.839603] env[61985]: INFO nova.compute.manager [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Took 1.16 seconds to destroy the instance on the hypervisor. [ 896.839872] env[61985]: DEBUG oslo.service.loopingcall [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.840555] env[61985]: DEBUG nova.compute.manager [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 896.840693] env[61985]: DEBUG nova.network.neutron [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 896.985868] env[61985]: DEBUG nova.objects.base [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Object Instance<5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf> lazy-loaded attributes: flavor,pci_requests {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 896.986186] env[61985]: DEBUG nova.network.neutron [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 897.013795] env[61985]: INFO nova.compute.manager [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Took 43.30 seconds to build instance. [ 897.039466] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 897.040426] env[61985]: ERROR nova.scheduler.client.report [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] [req-a5d914dd-dee3-40d7-9fff-a79bfcc02a78] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a5d914dd-dee3-40d7-9fff-a79bfcc02a78"}]} [ 897.040857] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ffa43bfa-8293-4a5f-904c-47441f8fbd63 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.054034] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 897.054034] env[61985]: value = "task-936026" [ 897.054034] env[61985]: _type = "Task" [ 897.054034] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.063592] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936026, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.066083] env[61985]: DEBUG nova.policy [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a78dbc667d24c819e6447740fc2d8e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2dbceebf07c43ac90ed80f8a09294b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 897.069548] env[61985]: DEBUG nova.scheduler.client.report [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 897.086023] env[61985]: DEBUG nova.scheduler.client.report [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 897.086277] env[61985]: DEBUG nova.compute.provider_tree [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 897.102796] env[61985]: DEBUG nova.scheduler.client.report [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 897.124203] env[61985]: DEBUG nova.scheduler.client.report [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 897.127339] env[61985]: DEBUG oslo_concurrency.lockutils [req-f1f11c7a-aed2-4a45-946d-8a4731a89a79 req-a4f71a74-7ee9-4751-99ca-350063b0aaa3 service nova] Releasing lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.128192] env[61985]: DEBUG oslo_concurrency.lockutils [req-5738acbc-f293-4997-91ea-9d75576bc756 req-c2a0e4ed-f77b-4923-b586-de4a82eccf1f service nova] Acquired lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.128506] env[61985]: DEBUG nova.network.neutron [req-5738acbc-f293-4997-91ea-9d75576bc756 req-c2a0e4ed-f77b-4923-b586-de4a82eccf1f service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Refreshing network info cache for port 885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 897.200515] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936025, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.306877] env[61985]: INFO nova.compute.manager [-] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Took 1.31 seconds to deallocate network for instance. [ 897.508647] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5f8287-ff1d-4047-860b-7375e8afe7d4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.516232] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42c3aa9c-316d-4772-ba88-7728700993af tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.993s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.517373] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d74b9bc-9626-400d-ab92-9de6ac2dbe0c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.550513] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2169a71-2f9b-44a5-aa9f-a6a10ac4c118 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.561444] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9dec9c-403e-43be-ae82-e7b48b48a379 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.568438] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936026, 'name': PowerOffVM_Task, 'duration_secs': 0.241742} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.569208] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 897.569454] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 897.570256] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bee345b-b655-4c0d-9003-8adaab8e3e4c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.582274] env[61985]: DEBUG nova.compute.provider_tree [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.588021] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 897.588298] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9e53339-aa93-45f6-9abc-72ab24ad25bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.697907] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936025, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.709576] env[61985]: DEBUG nova.network.neutron [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.803489] env[61985]: INFO nova.network.neutron [req-5738acbc-f293-4997-91ea-9d75576bc756 req-c2a0e4ed-f77b-4923-b586-de4a82eccf1f service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Port 885cda89-b90d-499a-a3b1-7155c07a1ce7 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 897.803778] env[61985]: DEBUG nova.network.neutron [req-5738acbc-f293-4997-91ea-9d75576bc756 req-c2a0e4ed-f77b-4923-b586-de4a82eccf1f service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.813311] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.086032] env[61985]: DEBUG nova.scheduler.client.report [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.199264] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936025, 'name': CloneVM_Task} progress is 95%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.212170] env[61985]: INFO nova.compute.manager [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Took 1.37 seconds to deallocate network for instance. [ 898.306950] env[61985]: DEBUG oslo_concurrency.lockutils [req-5738acbc-f293-4997-91ea-9d75576bc756 req-c2a0e4ed-f77b-4923-b586-de4a82eccf1f service nova] Releasing lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.410126] env[61985]: DEBUG nova.compute.manager [req-3700f7cf-ba2a-4d87-b7ee-5a80b2c8e9d1 req-8b633880-53a2-43b9-a261-187243950efd service nova] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Received event network-vif-deleted-e25b97ce-77e2-472d-a7a2-42899e22aab2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 898.593849] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.123s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.597439] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.097s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.600989] env[61985]: INFO nova.compute.claims [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.624209] env[61985]: DEBUG nova.network.neutron [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Successfully updated port: b464350a-1222-4ab5-9caf-df714dc8f763 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.627220] env[61985]: INFO nova.scheduler.client.report [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Deleted allocations for instance 4a94a6f9-1c86-4628-aa63-341f2c114e2a [ 898.700281] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936025, 'name': CloneVM_Task, 'duration_secs': 1.787128} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.700482] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Created linked-clone VM from snapshot [ 898.701458] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c246b86b-c1a4-40f6-ac0f-102b0c8177ea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.708719] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "488d0b95-97a9-4193-af62-cc92caf99625" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.709212] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "488d0b95-97a9-4193-af62-cc92caf99625" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.714456] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Uploading image dd40ba10-dabf-44e8-92c4-18c47ece2b25 {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 898.718845] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.738891] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 898.738891] env[61985]: value = "vm-211395" [ 898.738891] env[61985]: _type = "VirtualMachine" [ 898.738891] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 898.739207] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a6d3a100-e68c-43ac-95a2-7b8a1acf0515 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.746176] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lease: (returnval){ [ 898.746176] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5247030c-4794-e6a1-2ca1-637c5df227e5" [ 898.746176] env[61985]: _type = "HttpNfcLease" [ 898.746176] env[61985]: } obtained for exporting VM: (result){ [ 898.746176] env[61985]: value = "vm-211395" [ 898.746176] env[61985]: _type = "VirtualMachine" [ 898.746176] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 898.746465] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the lease: (returnval){ [ 898.746465] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5247030c-4794-e6a1-2ca1-637c5df227e5" [ 898.746465] env[61985]: _type = "HttpNfcLease" [ 898.746465] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 898.752923] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 898.752923] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5247030c-4794-e6a1-2ca1-637c5df227e5" [ 898.752923] env[61985]: _type = "HttpNfcLease" [ 898.752923] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 898.964042] env[61985]: DEBUG nova.compute.manager [req-de62e7b1-635a-4459-9c4f-142c3ef048b1 req-f617999f-3236-4950-a03b-280c37d8608e service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Received event network-changed-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 898.964042] env[61985]: DEBUG nova.compute.manager [req-de62e7b1-635a-4459-9c4f-142c3ef048b1 req-f617999f-3236-4950-a03b-280c37d8608e service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Refreshing instance network info cache due to event network-changed-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 898.964356] env[61985]: DEBUG oslo_concurrency.lockutils [req-de62e7b1-635a-4459-9c4f-142c3ef048b1 req-f617999f-3236-4950-a03b-280c37d8608e service nova] Acquiring lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.964356] env[61985]: DEBUG oslo_concurrency.lockutils [req-de62e7b1-635a-4459-9c4f-142c3ef048b1 req-f617999f-3236-4950-a03b-280c37d8608e service nova] Acquired lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.964501] env[61985]: DEBUG nova.network.neutron [req-de62e7b1-635a-4459-9c4f-142c3ef048b1 req-f617999f-3236-4950-a03b-280c37d8608e service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Refreshing network info cache for port 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 899.135419] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.135609] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.135796] env[61985]: DEBUG nova.network.neutron [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 899.138454] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9fa2845f-e544-4c31-9add-03be01fc6ec1 tempest-ServersTestJSON-1093001215 tempest-ServersTestJSON-1093001215-project-member] Lock "4a94a6f9-1c86-4628-aa63-341f2c114e2a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.425s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.217761] env[61985]: DEBUG nova.compute.manager [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 899.257151] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 899.257151] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5247030c-4794-e6a1-2ca1-637c5df227e5" [ 899.257151] env[61985]: _type = "HttpNfcLease" [ 899.257151] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 899.257489] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 899.257489] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5247030c-4794-e6a1-2ca1-637c5df227e5" [ 899.257489] env[61985]: _type = "HttpNfcLease" [ 899.257489] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 899.258404] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0342a1c-3686-4e1a-8ba8-b65ec3e2117f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.266636] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2bd29-08cf-595f-be0e-f76129a26886/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 899.266803] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2bd29-08cf-595f-be0e-f76129a26886/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 899.425126] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-41f1ac3f-5149-48cc-9ff7-cb6b76d188bc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.693293] env[61985]: WARNING nova.network.neutron [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] 5014dbd4-285f-43e0-ad95-2107c34e808a already exists in list: networks containing: ['5014dbd4-285f-43e0-ad95-2107c34e808a']. ignoring it [ 899.693519] env[61985]: WARNING nova.network.neutron [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] 5014dbd4-285f-43e0-ad95-2107c34e808a already exists in list: networks containing: ['5014dbd4-285f-43e0-ad95-2107c34e808a']. ignoring it [ 899.742737] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.753263] env[61985]: DEBUG nova.network.neutron [req-de62e7b1-635a-4459-9c4f-142c3ef048b1 req-f617999f-3236-4950-a03b-280c37d8608e service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updated VIF entry in instance network info cache for port 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 899.753682] env[61985]: DEBUG nova.network.neutron [req-de62e7b1-635a-4459-9c4f-142c3ef048b1 req-f617999f-3236-4950-a03b-280c37d8608e service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updating instance_info_cache with network_info: [{"id": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "address": "fa:16:3e:d4:ac:00", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60c2bc37-31", "ovs_interfaceid": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.019371] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf185af-24f2-4cc6-8871-82d1fec70bd3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.035079] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c292ae2b-fc0e-4347-8812-8256e0a547e4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.039452] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 900.039838] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 900.040282] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleting the datastore file [datastore1] e68cc5ff-83b7-4602-be50-ee4b1fa9ac65 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.040691] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7c0bb1e-4060-4df9-9e65-4d3a7bd0d606 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.081604] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d34679e-f504-456b-a115-da30302463d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.084561] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 900.084561] env[61985]: value = "task-936029" [ 900.084561] env[61985]: _type = "Task" [ 900.084561] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.093361] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a3a827-6ff9-4ae4-9294-3c8ab8add1ed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.102070] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936029, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.113521] env[61985]: DEBUG nova.compute.provider_tree [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.256961] env[61985]: DEBUG oslo_concurrency.lockutils [req-de62e7b1-635a-4459-9c4f-142c3ef048b1 req-f617999f-3236-4950-a03b-280c37d8608e service nova] Releasing lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.257659] env[61985]: DEBUG nova.compute.manager [req-de62e7b1-635a-4459-9c4f-142c3ef048b1 req-f617999f-3236-4950-a03b-280c37d8608e service nova] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Received event network-vif-deleted-885cda89-b90d-499a-a3b1-7155c07a1ce7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 900.512776] env[61985]: DEBUG nova.network.neutron [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updating instance_info_cache with network_info: [{"id": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "address": "fa:16:3e:68:94:dd", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4959d778-77", "ovs_interfaceid": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "94987e66-83ff-441e-924c-a937a0c6dc27", "address": "fa:16:3e:40:92:00", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94987e66-83", "ovs_interfaceid": "94987e66-83ff-441e-924c-a937a0c6dc27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b464350a-1222-4ab5-9caf-df714dc8f763", "address": "fa:16:3e:2e:99:64", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb464350a-12", "ovs_interfaceid": "b464350a-1222-4ab5-9caf-df714dc8f763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.594900] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936029, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.325786} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.595993] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.595993] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 900.595993] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 900.617070] env[61985]: DEBUG nova.scheduler.client.report [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.739288] env[61985]: DEBUG nova.compute.manager [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Stashing vm_state: stopped {{(pid=61985) _prep_resize /opt/stack/nova/nova/compute/manager.py:5630}} [ 901.016779] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.017565] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.017742] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.018695] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c747013d-655a-43f9-8b46-0001b510012a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.041441] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.041836] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.041836] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.042308] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.042308] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.042513] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.042790] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.043016] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.043217] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.043482] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.043689] env[61985]: DEBUG nova.virt.hardware [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.050506] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Reconfiguring VM to attach interface {{(pid=61985) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 901.051050] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-400b8646-8ede-4a31-813f-c5968682d67f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.070244] env[61985]: DEBUG oslo_vmware.api [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 901.070244] env[61985]: value = "task-936030" [ 901.070244] env[61985]: _type = "Task" [ 901.070244] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.078810] env[61985]: DEBUG oslo_vmware.api [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936030, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.127052] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.127699] env[61985]: DEBUG nova.compute.manager [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 901.130833] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.389s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.131939] env[61985]: INFO nova.compute.claims [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.261410] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.473443] env[61985]: DEBUG nova.compute.manager [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received event network-vif-plugged-b464350a-1222-4ab5-9caf-df714dc8f763 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 901.473443] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] Acquiring lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.473443] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.473691] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.473883] env[61985]: DEBUG nova.compute.manager [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] No waiting events found dispatching network-vif-plugged-b464350a-1222-4ab5-9caf-df714dc8f763 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 901.474069] env[61985]: WARNING nova.compute.manager [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received unexpected event network-vif-plugged-b464350a-1222-4ab5-9caf-df714dc8f763 for instance with vm_state active and task_state None. [ 901.474241] env[61985]: DEBUG nova.compute.manager [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received event network-changed-b464350a-1222-4ab5-9caf-df714dc8f763 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 901.474402] env[61985]: DEBUG nova.compute.manager [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Refreshing instance network info cache due to event network-changed-b464350a-1222-4ab5-9caf-df714dc8f763. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 901.474597] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] Acquiring lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.474737] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] Acquired lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.474896] env[61985]: DEBUG nova.network.neutron [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Refreshing network info cache for port b464350a-1222-4ab5-9caf-df714dc8f763 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 901.580634] env[61985]: DEBUG oslo_vmware.api [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.628443] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.628875] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.629106] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.629352] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.629528] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.629693] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.629924] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.630107] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.630294] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.630467] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.630650] env[61985]: DEBUG nova.virt.hardware [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.631713] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f754f70-d225-4ec8-a717-26dbcce0e6a5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.638735] env[61985]: DEBUG nova.compute.utils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 901.640165] env[61985]: DEBUG nova.compute.manager [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 901.640387] env[61985]: DEBUG nova.network.neutron [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 901.649188] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d5596a-ef61-4422-af0b-6cce0ac432fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.665684] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:ba:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bff6c3a1-cc80-46ca-86c0-6dbb029edddb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5b4f0fea-8589-4efc-b9b3-b59de5621860', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.673812] env[61985]: DEBUG oslo.service.loopingcall [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.674698] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 901.674788] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7941b449-2bbb-4154-9743-bcc5c926b28e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.690885] env[61985]: DEBUG nova.policy [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b5abdace901430cb1e431c9a933161e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bdb5f76b9b24a7ba4800a88482847da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 901.702151] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.702151] env[61985]: value = "task-936031" [ 901.702151] env[61985]: _type = "Task" [ 901.702151] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.712690] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936031, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.761730] env[61985]: DEBUG nova.compute.manager [req-2115962c-7579-4c04-92ee-2a30cb756c61 req-e819e639-8640-4d59-9327-8c33195b6768 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Received event network-changed-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 901.762328] env[61985]: DEBUG nova.compute.manager [req-2115962c-7579-4c04-92ee-2a30cb756c61 req-e819e639-8640-4d59-9327-8c33195b6768 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Refreshing instance network info cache due to event network-changed-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 901.762695] env[61985]: DEBUG oslo_concurrency.lockutils [req-2115962c-7579-4c04-92ee-2a30cb756c61 req-e819e639-8640-4d59-9327-8c33195b6768 service nova] Acquiring lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.762961] env[61985]: DEBUG oslo_concurrency.lockutils [req-2115962c-7579-4c04-92ee-2a30cb756c61 req-e819e639-8640-4d59-9327-8c33195b6768 service nova] Acquired lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.763250] env[61985]: DEBUG nova.network.neutron [req-2115962c-7579-4c04-92ee-2a30cb756c61 req-e819e639-8640-4d59-9327-8c33195b6768 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Refreshing network info cache for port 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.083735] env[61985]: DEBUG oslo_vmware.api [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936030, 'name': ReconfigVM_Task, 'duration_secs': 0.769442} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.084435] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.084784] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Reconfigured VM to attach interface {{(pid=61985) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 902.138546] env[61985]: DEBUG nova.network.neutron [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Successfully created port: a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.147016] env[61985]: DEBUG nova.compute.manager [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 902.197149] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "c64056d8-0c22-48db-bf4a-a60f468bdad0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.217025] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936031, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.382369] env[61985]: DEBUG nova.network.neutron [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updated VIF entry in instance network info cache for port b464350a-1222-4ab5-9caf-df714dc8f763. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 902.382878] env[61985]: DEBUG nova.network.neutron [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updating instance_info_cache with network_info: [{"id": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "address": "fa:16:3e:68:94:dd", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4959d778-77", "ovs_interfaceid": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "94987e66-83ff-441e-924c-a937a0c6dc27", "address": "fa:16:3e:40:92:00", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94987e66-83", "ovs_interfaceid": "94987e66-83ff-441e-924c-a937a0c6dc27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b464350a-1222-4ab5-9caf-df714dc8f763", "address": "fa:16:3e:2e:99:64", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb464350a-12", "ovs_interfaceid": "b464350a-1222-4ab5-9caf-df714dc8f763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.388018] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquiring lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.388018] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.388018] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquiring lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.388018] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.388018] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.390143] env[61985]: INFO nova.compute.manager [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Terminating instance [ 902.392080] env[61985]: DEBUG nova.compute.manager [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 902.392293] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 902.394281] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089d6b15-c249-4e48-9c9c-7a2ab4b2aea1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.406041] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 902.409881] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8beddf0-a16a-4c59-8464-82d065a33abe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.419465] env[61985]: DEBUG oslo_vmware.api [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for the task: (returnval){ [ 902.419465] env[61985]: value = "task-936032" [ 902.419465] env[61985]: _type = "Task" [ 902.419465] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.429413] env[61985]: DEBUG oslo_vmware.api [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936032, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.553456] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b494408a-6727-4cc9-b282-f77d2864c6c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.563031] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4928e6-0846-47bf-91cd-b7afcd7ebbda {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.598085] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73d3481c-fab7-49cc-afc4-c2440ad59a87 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-b464350a-1222-4ab5-9caf-df714dc8f763" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.755s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.604645] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e026fc4d-ac51-4b49-b9d0-855b6aa344c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.618404] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e15d8b4-e639-4ebf-88ee-95fbff25a390 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.634661] env[61985]: DEBUG nova.compute.provider_tree [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.716153] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936031, 'name': CreateVM_Task, 'duration_secs': 0.679327} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.717058] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 902.717842] env[61985]: DEBUG oslo_concurrency.lockutils [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.718028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.718392] env[61985]: DEBUG oslo_concurrency.lockutils [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.718986] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-594439dd-2d39-433c-91b0-28f05c69d11f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.726284] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 902.726284] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5267efa8-8e66-af66-1262-ace8a4f9ea6c" [ 902.726284] env[61985]: _type = "Task" [ 902.726284] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.740330] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5267efa8-8e66-af66-1262-ace8a4f9ea6c, 'name': SearchDatastore_Task} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.740773] env[61985]: DEBUG oslo_concurrency.lockutils [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.740885] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.741268] env[61985]: DEBUG oslo_concurrency.lockutils [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.741386] env[61985]: DEBUG oslo_concurrency.lockutils [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.741550] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.741993] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-82071490-5120-4641-9e83-b37e8e6417ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.750970] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.751115] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 902.751940] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3db588b-6da2-494a-911a-2ac82b847784 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.758841] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 902.758841] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5293dd9b-8987-f0ea-e0f3-c0b5864520a9" [ 902.758841] env[61985]: _type = "Task" [ 902.758841] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.766804] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5293dd9b-8987-f0ea-e0f3-c0b5864520a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.816675] env[61985]: DEBUG nova.network.neutron [req-2115962c-7579-4c04-92ee-2a30cb756c61 req-e819e639-8640-4d59-9327-8c33195b6768 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updated VIF entry in instance network info cache for port 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 902.816675] env[61985]: DEBUG nova.network.neutron [req-2115962c-7579-4c04-92ee-2a30cb756c61 req-e819e639-8640-4d59-9327-8c33195b6768 service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updating instance_info_cache with network_info: [{"id": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "address": "fa:16:3e:d4:ac:00", "network": {"id": "c0793e67-29a1-4e6a-acc6-ff204b558504", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1639161399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d49f3925a4074d779d6dd68164545cfb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60c2bc37-31", "ovs_interfaceid": "60c2bc37-3190-4dfb-8b71-fd6eb3c3949b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.888190] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] Releasing lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.888190] env[61985]: DEBUG nova.compute.manager [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Received event network-changed-97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 902.888190] env[61985]: DEBUG nova.compute.manager [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Refreshing instance network info cache due to event network-changed-97f641fb-39ac-441c-bacc-6705d1ea6e98. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 902.888190] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] Acquiring lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.888190] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] Acquired lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.888190] env[61985]: DEBUG nova.network.neutron [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Refreshing network info cache for port 97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.929989] env[61985]: DEBUG oslo_vmware.api [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936032, 'name': PowerOffVM_Task, 'duration_secs': 0.225375} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.931281] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 902.931281] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 902.931281] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8ec98d5-728a-4709-a3cc-91da86c67699 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.994704] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 902.994868] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 902.995079] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Deleting the datastore file [datastore2] 54ba5cfd-185a-4c58-aa5e-83cc096a482e {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.995319] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b78d4e91-dac7-4b32-b9d6-33bb6ce22355 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.004239] env[61985]: DEBUG oslo_vmware.api [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for the task: (returnval){ [ 903.004239] env[61985]: value = "task-936034" [ 903.004239] env[61985]: _type = "Task" [ 903.004239] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.013847] env[61985]: DEBUG oslo_vmware.api [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936034, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.142231] env[61985]: DEBUG nova.scheduler.client.report [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.159780] env[61985]: DEBUG nova.compute.manager [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 903.199039] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 903.199039] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 903.199039] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.199039] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 903.199039] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.199039] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 903.199368] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 903.199600] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 903.199827] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 903.200198] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 903.200451] env[61985]: DEBUG nova.virt.hardware [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.201715] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83402397-fa65-430b-9ead-81e8b1be6382 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.212514] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c6a51b-accd-4165-884d-11e58e17a771 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.271179] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5293dd9b-8987-f0ea-e0f3-c0b5864520a9, 'name': SearchDatastore_Task, 'duration_secs': 0.011769} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.271179] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eabeb565-8e22-466a-a652-264b4a3a3cdf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.275178] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 903.275178] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d6347c-b614-facd-e1f7-260ef93ae6db" [ 903.275178] env[61985]: _type = "Task" [ 903.275178] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.285847] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d6347c-b614-facd-e1f7-260ef93ae6db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.319129] env[61985]: DEBUG oslo_concurrency.lockutils [req-2115962c-7579-4c04-92ee-2a30cb756c61 req-e819e639-8640-4d59-9327-8c33195b6768 service nova] Releasing lock "refresh_cache-3694e20c-ce37-4097-9991-8a06f38b2734" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.514362] env[61985]: DEBUG oslo_vmware.api [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Task: {'id': task-936034, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209107} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.514944] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.515248] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 903.515470] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 903.515654] env[61985]: INFO nova.compute.manager [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 903.515968] env[61985]: DEBUG oslo.service.loopingcall [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.516171] env[61985]: DEBUG nova.compute.manager [-] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 903.516267] env[61985]: DEBUG nova.network.neutron [-] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 903.648885] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.649561] env[61985]: DEBUG nova.compute.manager [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 903.657020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 33.125s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.791023] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d6347c-b614-facd-e1f7-260ef93ae6db, 'name': SearchDatastore_Task, 'duration_secs': 0.015497} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.791479] env[61985]: DEBUG oslo_concurrency.lockutils [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.791795] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] e68cc5ff-83b7-4602-be50-ee4b1fa9ac65/e68cc5ff-83b7-4602-be50-ee4b1fa9ac65.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 903.792305] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07efaf68-0116-45fe-bdce-e5c63213163e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.799831] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 903.799831] env[61985]: value = "task-936035" [ 903.799831] env[61985]: _type = "Task" [ 903.799831] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.809291] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936035, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.935097] env[61985]: DEBUG nova.network.neutron [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updated VIF entry in instance network info cache for port 97f641fb-39ac-441c-bacc-6705d1ea6e98. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 903.935550] env[61985]: DEBUG nova.network.neutron [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updating instance_info_cache with network_info: [{"id": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "address": "fa:16:3e:fc:45:f4", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97f641fb-39", "ovs_interfaceid": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.133345] env[61985]: DEBUG nova.compute.manager [req-d2a4ae68-783b-4891-a585-4a60fdc70ea2 req-c6b6d5db-8ac5-4897-bbb8-1aa1608aa7c2 service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Received event network-vif-deleted-8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 904.133345] env[61985]: INFO nova.compute.manager [req-d2a4ae68-783b-4891-a585-4a60fdc70ea2 req-c6b6d5db-8ac5-4897-bbb8-1aa1608aa7c2 service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Neutron deleted interface 8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4; detaching it from the instance and deleting it from the info cache [ 904.133345] env[61985]: DEBUG nova.network.neutron [req-d2a4ae68-783b-4891-a585-4a60fdc70ea2 req-c6b6d5db-8ac5-4897-bbb8-1aa1608aa7c2 service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.169984] env[61985]: DEBUG nova.compute.utils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.178460] env[61985]: DEBUG nova.compute.manager [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 904.179035] env[61985]: DEBUG nova.network.neutron [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 904.240330] env[61985]: DEBUG nova.policy [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faa19f385cd0426d8565e31e286800d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5625013599b44418bd56eb604e14be58', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.301777] env[61985]: DEBUG nova.compute.manager [req-0e22532d-1214-4397-bacf-1dd0928704ad req-7e231fb0-5de0-4923-84c9-f4300e8d18fc service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Received event network-vif-plugged-a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 904.305028] env[61985]: DEBUG oslo_concurrency.lockutils [req-0e22532d-1214-4397-bacf-1dd0928704ad req-7e231fb0-5de0-4923-84c9-f4300e8d18fc service nova] Acquiring lock "073a7668-39e6-480d-9350-835a0282b456-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.305028] env[61985]: DEBUG oslo_concurrency.lockutils [req-0e22532d-1214-4397-bacf-1dd0928704ad req-7e231fb0-5de0-4923-84c9-f4300e8d18fc service nova] Lock "073a7668-39e6-480d-9350-835a0282b456-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.305028] env[61985]: DEBUG oslo_concurrency.lockutils [req-0e22532d-1214-4397-bacf-1dd0928704ad req-7e231fb0-5de0-4923-84c9-f4300e8d18fc service nova] Lock "073a7668-39e6-480d-9350-835a0282b456-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.305028] env[61985]: DEBUG nova.compute.manager [req-0e22532d-1214-4397-bacf-1dd0928704ad req-7e231fb0-5de0-4923-84c9-f4300e8d18fc service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] No waiting events found dispatching network-vif-plugged-a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 904.305028] env[61985]: WARNING nova.compute.manager [req-0e22532d-1214-4397-bacf-1dd0928704ad req-7e231fb0-5de0-4923-84c9-f4300e8d18fc service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Received unexpected event network-vif-plugged-a5e0f361-c876-4212-8c4a-dfdfb17d5d4f for instance with vm_state building and task_state spawning. [ 904.322464] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936035, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.439392] env[61985]: DEBUG oslo_concurrency.lockutils [req-d4f36f83-2113-4247-8cad-bf0f9b7379c5 req-151e8f25-7c89-4927-9316-82b8ada6cdfc service nova] Releasing lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.450715] env[61985]: DEBUG nova.network.neutron [-] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.532398] env[61985]: DEBUG nova.network.neutron [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Successfully created port: f9834461-44c6-4878-8da9-e4a918ad3f51 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.638698] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3da60301-ac7f-4fde-9120-3474183d508c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.648661] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654bb342-071d-4cce-ba73-9766056e201e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.682942] env[61985]: DEBUG nova.compute.manager [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 904.687535] env[61985]: INFO nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating resource usage from migration 9a221221-2558-443c-8a47-3d0ca57efb01 [ 904.691108] env[61985]: DEBUG nova.compute.manager [req-d2a4ae68-783b-4891-a585-4a60fdc70ea2 req-c6b6d5db-8ac5-4897-bbb8-1aa1608aa7c2 service nova] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Detach interface failed, port_id=8be34cd9-0c58-4d58-9c2f-21e98ea0b3f4, reason: Instance 54ba5cfd-185a-4c58-aa5e-83cc096a482e could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 904.717253] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 3694e20c-ce37-4097-9991-8a06f38b2734 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.717330] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 761d5954-a5ca-4459-a1d6-bfc59b284bf4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.717459] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance cfd59b61-cca9-48d5-85e1-1f45d13f1e88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.717781] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.717781] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 1f284789-1e7b-4e9f-9670-34e8e25cd797 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.717906] env[61985]: WARNING nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 17bdf8ec-13d2-459c-bc8e-db6a274fc27e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 904.717964] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 59e742d9-c51a-47dd-9ef2-2f070eae94c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.718087] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance ce757dda-f58a-47b3-b319-e6b01f05b20b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.718210] env[61985]: WARNING nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 5b3a8ba3-48cb-472a-ab45-4e51af68812c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 904.718536] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance fea0cf39-e851-409f-86f5-31cc128a44dc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.718536] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance with task_state "deleting" is not being actively managed by this compute host but has allocations referencing this compute node (aed7e5b3-c662-4538-8447-c4f67b460215): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 904.718670] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance e68cc5ff-83b7-4602-be50-ee4b1fa9ac65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.718707] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 518635bf-73ee-404b-ae6a-dc4ee23009d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.719226] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 54ba5cfd-185a-4c58-aa5e-83cc096a482e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.719226] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 6dd8218b-c14d-40c9-87df-097fab06c669 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.719226] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 073a7668-39e6-480d-9350-835a0282b456 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.722616] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 904.758794] env[61985]: DEBUG nova.network.neutron [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Successfully updated port: a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.784963] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-94987e66-83ff-441e-924c-a937a0c6dc27" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.785247] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-94987e66-83ff-441e-924c-a937a0c6dc27" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.790777] env[61985]: DEBUG nova.compute.manager [req-b27fd0ee-e253-47a7-8cc8-01733dc4133c req-fd793a38-e88b-4fe3-8d73-b7806f8479f8 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Received event network-changed-a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 904.792594] env[61985]: DEBUG nova.compute.manager [req-b27fd0ee-e253-47a7-8cc8-01733dc4133c req-fd793a38-e88b-4fe3-8d73-b7806f8479f8 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Refreshing instance network info cache due to event network-changed-a5e0f361-c876-4212-8c4a-dfdfb17d5d4f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 904.792594] env[61985]: DEBUG oslo_concurrency.lockutils [req-b27fd0ee-e253-47a7-8cc8-01733dc4133c req-fd793a38-e88b-4fe3-8d73-b7806f8479f8 service nova] Acquiring lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.792594] env[61985]: DEBUG oslo_concurrency.lockutils [req-b27fd0ee-e253-47a7-8cc8-01733dc4133c req-fd793a38-e88b-4fe3-8d73-b7806f8479f8 service nova] Acquired lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.792594] env[61985]: DEBUG nova.network.neutron [req-b27fd0ee-e253-47a7-8cc8-01733dc4133c req-fd793a38-e88b-4fe3-8d73-b7806f8479f8 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Refreshing network info cache for port a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.813294] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936035, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.559744} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.813625] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] e68cc5ff-83b7-4602-be50-ee4b1fa9ac65/e68cc5ff-83b7-4602-be50-ee4b1fa9ac65.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 904.813815] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.814102] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3c09a776-f933-4e87-bbdc-91e4ec3a3fc8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.822455] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 904.822455] env[61985]: value = "task-936036" [ 904.822455] env[61985]: _type = "Task" [ 904.822455] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.832243] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936036, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.952435] env[61985]: INFO nova.compute.manager [-] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Took 1.44 seconds to deallocate network for instance. [ 905.226336] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 986c576e-be02-48ac-b24c-72edccab25be has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 905.262865] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.289713] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.290396] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.290837] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c86873a-5530-4981-b9be-ba75a93789b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.314624] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c948622c-61c1-4e7c-8bad-d6d9c9ef47fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.345018] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Reconfiguring VM to detach interface {{(pid=61985) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 905.347892] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3984f6b6-b12e-4c2f-9ee0-de78b11c7533 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.361490] env[61985]: DEBUG nova.network.neutron [req-b27fd0ee-e253-47a7-8cc8-01733dc4133c req-fd793a38-e88b-4fe3-8d73-b7806f8479f8 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 905.369040] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936036, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065804} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.370212] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.370585] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 905.370585] env[61985]: value = "task-936037" [ 905.370585] env[61985]: _type = "Task" [ 905.370585] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.371289] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11f7948-02c1-460a-b200-dd8171619570 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.400021] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] e68cc5ff-83b7-4602-be50-ee4b1fa9ac65/e68cc5ff-83b7-4602-be50-ee4b1fa9ac65.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.400425] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.400683] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-843ba6f5-3d2a-4207-b0a4-f6f205f76081 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.422517] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 905.422517] env[61985]: value = "task-936038" [ 905.422517] env[61985]: _type = "Task" [ 905.422517] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.431095] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936038, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.459547] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.497629] env[61985]: DEBUG nova.network.neutron [req-b27fd0ee-e253-47a7-8cc8-01733dc4133c req-fd793a38-e88b-4fe3-8d73-b7806f8479f8 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.698789] env[61985]: DEBUG nova.compute.manager [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 905.723582] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.723845] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.724019] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.724307] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.724482] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.724637] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.724866] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.725388] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.725388] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.725388] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.725606] env[61985]: DEBUG nova.virt.hardware [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.726663] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d08888a-db65-4198-9625-f2d3c51af784 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.730987] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance a77ca23f-b2c0-4822-8e48-3e47e0dadb27 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 905.736428] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06452597-fdc2-4cb4-94fd-0d34d1ab62cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.901752] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.939636] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936038, 'name': ReconfigVM_Task, 'duration_secs': 0.380999} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.940128] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Reconfigured VM instance instance-00000040 to attach disk [datastore2] e68cc5ff-83b7-4602-be50-ee4b1fa9ac65/e68cc5ff-83b7-4602-be50-ee4b1fa9ac65.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.940888] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f77a5fc7-79e9-4c35-b370-7d32830ca92c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.951262] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 905.951262] env[61985]: value = "task-936039" [ 905.951262] env[61985]: _type = "Task" [ 905.951262] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.962320] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936039, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.001096] env[61985]: DEBUG oslo_concurrency.lockutils [req-b27fd0ee-e253-47a7-8cc8-01733dc4133c req-fd793a38-e88b-4fe3-8d73-b7806f8479f8 service nova] Releasing lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.001251] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.002008] env[61985]: DEBUG nova.network.neutron [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 906.236461] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 87006fbf-ea90-4d9a-88af-001de424ac14 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 906.384567] env[61985]: DEBUG nova.compute.manager [req-18210079-1f1e-4172-94aa-d3314a45973a req-21802ff2-0893-4dbc-ae2e-e39c975edae1 service nova] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Received event network-vif-plugged-f9834461-44c6-4878-8da9-e4a918ad3f51 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 906.384777] env[61985]: DEBUG oslo_concurrency.lockutils [req-18210079-1f1e-4172-94aa-d3314a45973a req-21802ff2-0893-4dbc-ae2e-e39c975edae1 service nova] Acquiring lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.385026] env[61985]: DEBUG oslo_concurrency.lockutils [req-18210079-1f1e-4172-94aa-d3314a45973a req-21802ff2-0893-4dbc-ae2e-e39c975edae1 service nova] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.385214] env[61985]: DEBUG oslo_concurrency.lockutils [req-18210079-1f1e-4172-94aa-d3314a45973a req-21802ff2-0893-4dbc-ae2e-e39c975edae1 service nova] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.385391] env[61985]: DEBUG nova.compute.manager [req-18210079-1f1e-4172-94aa-d3314a45973a req-21802ff2-0893-4dbc-ae2e-e39c975edae1 service nova] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] No waiting events found dispatching network-vif-plugged-f9834461-44c6-4878-8da9-e4a918ad3f51 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.385571] env[61985]: WARNING nova.compute.manager [req-18210079-1f1e-4172-94aa-d3314a45973a req-21802ff2-0893-4dbc-ae2e-e39c975edae1 service nova] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Received unexpected event network-vif-plugged-f9834461-44c6-4878-8da9-e4a918ad3f51 for instance with vm_state building and task_state spawning. [ 906.389795] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.444874] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.445159] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.461761] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936039, 'name': Rename_Task, 'duration_secs': 0.198036} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.461761] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 906.461761] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e7a6964-4510-4663-8eda-a4abdcf482c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.468946] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 906.468946] env[61985]: value = "task-936040" [ 906.468946] env[61985]: _type = "Task" [ 906.468946] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.477838] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936040, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.544074] env[61985]: DEBUG nova.network.neutron [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 906.598125] env[61985]: DEBUG nova.network.neutron [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Successfully updated port: f9834461-44c6-4878-8da9-e4a918ad3f51 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.739283] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 5c2a7072-5b14-4ec7-8060-4e21eea927e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 906.771988] env[61985]: DEBUG nova.network.neutron [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Updating instance_info_cache with network_info: [{"id": "a5e0f361-c876-4212-8c4a-dfdfb17d5d4f", "address": "fa:16:3e:3a:9a:10", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5e0f361-c8", "ovs_interfaceid": "a5e0f361-c876-4212-8c4a-dfdfb17d5d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.827053] env[61985]: DEBUG nova.compute.manager [req-5268e33b-94cd-4f2a-93db-01e79229a029 req-d20ab4f6-775c-49ac-abc6-affb9da7265b service nova] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Received event network-changed-f9834461-44c6-4878-8da9-e4a918ad3f51 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 906.827507] env[61985]: DEBUG nova.compute.manager [req-5268e33b-94cd-4f2a-93db-01e79229a029 req-d20ab4f6-775c-49ac-abc6-affb9da7265b service nova] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Refreshing instance network info cache due to event network-changed-f9834461-44c6-4878-8da9-e4a918ad3f51. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 906.827507] env[61985]: DEBUG oslo_concurrency.lockutils [req-5268e33b-94cd-4f2a-93db-01e79229a029 req-d20ab4f6-775c-49ac-abc6-affb9da7265b service nova] Acquiring lock "refresh_cache-aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.827829] env[61985]: DEBUG oslo_concurrency.lockutils [req-5268e33b-94cd-4f2a-93db-01e79229a029 req-d20ab4f6-775c-49ac-abc6-affb9da7265b service nova] Acquired lock "refresh_cache-aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.827866] env[61985]: DEBUG nova.network.neutron [req-5268e33b-94cd-4f2a-93db-01e79229a029 req-d20ab4f6-775c-49ac-abc6-affb9da7265b service nova] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Refreshing network info cache for port f9834461-44c6-4878-8da9-e4a918ad3f51 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 906.887560] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.948690] env[61985]: DEBUG nova.compute.utils [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 906.979722] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936040, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.100925] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "refresh_cache-aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.242937] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 62824eac-9412-466a-abcf-1010f6a829e4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 907.275090] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.275445] env[61985]: DEBUG nova.compute.manager [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Instance network_info: |[{"id": "a5e0f361-c876-4212-8c4a-dfdfb17d5d4f", "address": "fa:16:3e:3a:9a:10", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5e0f361-c8", "ovs_interfaceid": "a5e0f361-c876-4212-8c4a-dfdfb17d5d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 907.276046] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:9a:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5e0f361-c876-4212-8c4a-dfdfb17d5d4f', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.284658] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Creating folder: Project (7bdb5f76b9b24a7ba4800a88482847da). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 907.284983] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-99094fee-431a-4869-8c0b-cf5af9873185 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.296995] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Created folder: Project (7bdb5f76b9b24a7ba4800a88482847da) in parent group-v211285. [ 907.297231] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Creating folder: Instances. Parent ref: group-v211397. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 907.297475] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fd1764f2-8b13-43dc-a7db-16933fbe9e21 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.306045] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Created folder: Instances in parent group-v211397. [ 907.306292] env[61985]: DEBUG oslo.service.loopingcall [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.306492] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 073a7668-39e6-480d-9350-835a0282b456] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 907.306711] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9040c057-91e2-4fc7-9b50-35418c587020 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.325742] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 907.325742] env[61985]: value = "task-936043" [ 907.325742] env[61985]: _type = "Task" [ 907.325742] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.334724] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936043, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.378078] env[61985]: DEBUG nova.network.neutron [req-5268e33b-94cd-4f2a-93db-01e79229a029 req-d20ab4f6-775c-49ac-abc6-affb9da7265b service nova] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 907.389916] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.452098] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.477426] env[61985]: DEBUG nova.network.neutron [req-5268e33b-94cd-4f2a-93db-01e79229a029 req-d20ab4f6-775c-49ac-abc6-affb9da7265b service nova] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.484574] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936040, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.746302] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance cd8e2eaf-3c13-452b-b2e6-8107f0219378 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 907.836567] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936043, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.890860] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.980252] env[61985]: DEBUG oslo_vmware.api [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936040, 'name': PowerOnVM_Task, 'duration_secs': 1.267127} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.980544] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 907.980795] env[61985]: DEBUG nova.compute.manager [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.981550] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e98501-2ad3-4c11-b34a-5de6cccfb900 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.985906] env[61985]: DEBUG oslo_concurrency.lockutils [req-5268e33b-94cd-4f2a-93db-01e79229a029 req-d20ab4f6-775c-49ac-abc6-affb9da7265b service nova] Releasing lock "refresh_cache-aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.986572] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "refresh_cache-aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.986871] env[61985]: DEBUG nova.network.neutron [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 908.249519] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 488d0b95-97a9-4193-af62-cc92caf99625 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 908.249910] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Migration 9a221221-2558-443c-8a47-3d0ca57efb01 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 908.249910] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 2322ab72-9841-41fb-9d60-2812baabe108 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 908.250125] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 15 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 908.250276] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3456MB phys_disk=200GB used_disk=15GB total_vcpus=48 used_vcpus=15 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 908.340574] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936043, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.390713] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.508466] env[61985]: DEBUG oslo_concurrency.lockutils [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.523356] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.524020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.524020] env[61985]: INFO nova.compute.manager [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Attaching volume 2e7a203f-be54-4d0d-b6df-a5b2894cd4d4 to /dev/sdb [ 908.529410] env[61985]: DEBUG nova.network.neutron [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 908.576024] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aee8cf9-e15d-4303-9540-b4783f8adf07 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.584456] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ffa100-ebbc-4f07-a3f6-63127de4aef4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.599584] env[61985]: DEBUG nova.virt.block_device [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Updating existing volume attachment record: 3d65dd01-2196-4988-a107-84e82828bac9 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 908.621454] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5dac5d1-0881-4f64-94b8-253343eee7c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.628980] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc8ce9a-6395-44e8-994b-58bb6ca489e6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.660209] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85ae15b-a5b9-444f-b1ab-84ec27cef2f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.668564] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc20175c-1363-462c-a73b-de67028ed44d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.681758] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.781981] env[61985]: DEBUG nova.network.neutron [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Updating instance_info_cache with network_info: [{"id": "f9834461-44c6-4878-8da9-e4a918ad3f51", "address": "fa:16:3e:0b:28:40", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9834461-44", "ovs_interfaceid": "f9834461-44c6-4878-8da9-e4a918ad3f51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.842684] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936043, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.892267] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.040890] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.042332] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.042332] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.042332] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.042332] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.044245] env[61985]: INFO nova.compute.manager [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Terminating instance [ 909.046469] env[61985]: DEBUG nova.compute.manager [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 909.046758] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 909.047736] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12aa83e1-74fa-4d46-aa47-830b5f5f8771 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.054884] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 909.055249] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2412284-8fb3-4bb8-ab0f-077f42890d95 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.061378] env[61985]: DEBUG oslo_vmware.api [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 909.061378] env[61985]: value = "task-936047" [ 909.061378] env[61985]: _type = "Task" [ 909.061378] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.071947] env[61985]: DEBUG oslo_vmware.api [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936047, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.186153] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.199574] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2bd29-08cf-595f-be0e-f76129a26886/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 909.200630] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50dbf997-fe5c-40ac-a160-51771c12da21 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.207482] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2bd29-08cf-595f-be0e-f76129a26886/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 909.208286] env[61985]: ERROR oslo_vmware.rw_handles [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2bd29-08cf-595f-be0e-f76129a26886/disk-0.vmdk due to incomplete transfer. [ 909.209249] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cf3a6acb-2d38-4971-82f0-aeb509614e56 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.216657] env[61985]: DEBUG oslo_vmware.rw_handles [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b2bd29-08cf-595f-be0e-f76129a26886/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 909.216885] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Uploaded image dd40ba10-dabf-44e8-92c4-18c47ece2b25 to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 909.219627] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 909.220196] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f6e85954-582d-4816-a169-9f4050eaf61a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.226047] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 909.226047] env[61985]: value = "task-936048" [ 909.226047] env[61985]: _type = "Task" [ 909.226047] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.235145] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936048, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.287131] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "refresh_cache-aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.287454] env[61985]: DEBUG nova.compute.manager [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Instance network_info: |[{"id": "f9834461-44c6-4878-8da9-e4a918ad3f51", "address": "fa:16:3e:0b:28:40", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9834461-44", "ovs_interfaceid": "f9834461-44c6-4878-8da9-e4a918ad3f51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 909.287992] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:28:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7908211b-df93-467b-87a8-3c3d29b03de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9834461-44c6-4878-8da9-e4a918ad3f51', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 909.296318] env[61985]: DEBUG oslo.service.loopingcall [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.296493] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 909.296757] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ad6ca918-0ce9-4f81-bdbf-8e0c31f38cae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.317960] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 909.317960] env[61985]: value = "task-936049" [ 909.317960] env[61985]: _type = "Task" [ 909.317960] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.326461] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936049, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.337283] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936043, 'name': CreateVM_Task, 'duration_secs': 1.530057} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.337468] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 073a7668-39e6-480d-9350-835a0282b456] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 909.338225] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.338397] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.338733] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 909.338994] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee6c06fe-58e2-4eb2-9850-73b61d29b3ca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.344627] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 909.344627] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c68f70-1be1-1fd7-d457-83aa66307539" [ 909.344627] env[61985]: _type = "Task" [ 909.344627] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.352955] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c68f70-1be1-1fd7-d457-83aa66307539, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.395345] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.572798] env[61985]: DEBUG oslo_vmware.api [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936047, 'name': PowerOffVM_Task, 'duration_secs': 0.266284} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.573119] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 909.573390] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 909.573597] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6cbe86b9-473c-4eed-bb5b-97a61feb89ad {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.630788] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "625a3143-d138-4b52-aeb7-9e365c6f1128" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.631416] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.668519] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 909.668801] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 909.669052] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleting the datastore file [datastore2] e68cc5ff-83b7-4602-be50-ee4b1fa9ac65 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.669381] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28afe5f9-892b-483b-a00e-d5dc15ed176c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.676926] env[61985]: DEBUG oslo_vmware.api [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 909.676926] env[61985]: value = "task-936051" [ 909.676926] env[61985]: _type = "Task" [ 909.676926] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.687244] env[61985]: DEBUG oslo_vmware.api [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936051, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.691084] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 909.691470] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.038s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.691696] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.993s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.691921] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.694066] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.550s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.695662] env[61985]: INFO nova.compute.claims [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.718039] env[61985]: INFO nova.scheduler.client.report [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Deleted allocations for instance 17bdf8ec-13d2-459c-bc8e-db6a274fc27e [ 909.735662] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936048, 'name': Destroy_Task, 'duration_secs': 0.437169} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.735957] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Destroyed the VM [ 909.736302] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 909.736600] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0fca536d-7bc5-4850-af46-550810b2345e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.743373] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 909.743373] env[61985]: value = "task-936052" [ 909.743373] env[61985]: _type = "Task" [ 909.743373] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.752075] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936052, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.828831] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936049, 'name': CreateVM_Task, 'duration_secs': 0.36354} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.829045] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 909.830207] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.830207] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.830551] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 909.830804] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c129e955-6a0e-4805-b16d-4de08a5af12f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.836293] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 909.836293] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5279d3e7-edc4-ea5c-4fc6-e5ad20d30af7" [ 909.836293] env[61985]: _type = "Task" [ 909.836293] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.844804] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5279d3e7-edc4-ea5c-4fc6-e5ad20d30af7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.853770] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c68f70-1be1-1fd7-d457-83aa66307539, 'name': SearchDatastore_Task, 'duration_secs': 0.012187} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.854056] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.854299] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.854595] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.854679] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.854862] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.855126] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d835922-ef86-4feb-9220-6aafdec52ca8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.862521] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.862699] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 909.863382] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6099c569-ab66-4945-b9db-3a415c38725d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.867743] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 909.867743] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b2a11f-9fb4-6309-f7e9-2d6d6bcc8b1b" [ 909.867743] env[61985]: _type = "Task" [ 909.867743] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.875110] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b2a11f-9fb4-6309-f7e9-2d6d6bcc8b1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.891631] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.135026] env[61985]: DEBUG nova.compute.manager [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 910.186459] env[61985]: DEBUG oslo_vmware.api [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936051, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176681} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.186713] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.186900] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 910.187094] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 910.187275] env[61985]: INFO nova.compute.manager [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Took 1.14 seconds to destroy the instance on the hypervisor. [ 910.187516] env[61985]: DEBUG oslo.service.loopingcall [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.187730] env[61985]: DEBUG nova.compute.manager [-] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.187828] env[61985]: DEBUG nova.network.neutron [-] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 910.224853] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17a94376-0940-44dd-b9ff-04786083b5da tempest-ServerGroupTestJSON-1262090341 tempest-ServerGroupTestJSON-1262090341-project-member] Lock "17bdf8ec-13d2-459c-bc8e-db6a274fc27e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.595s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.252836] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936052, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.348282] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5279d3e7-edc4-ea5c-4fc6-e5ad20d30af7, 'name': SearchDatastore_Task, 'duration_secs': 0.015051} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.348620] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.348862] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.349117] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.349272] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.349457] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.349720] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fa4c631-7828-4e9a-8234-d08995fa4be4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.358602] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.358602] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 910.359323] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-902cef03-504a-4045-a353-f709e9aa1167 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.368019] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 910.368019] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f33f62-93ee-6848-1809-7f061328b62c" [ 910.368019] env[61985]: _type = "Task" [ 910.368019] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.381765] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f33f62-93ee-6848-1809-7f061328b62c, 'name': SearchDatastore_Task, 'duration_secs': 0.009029} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.388424] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b2a11f-9fb4-6309-f7e9-2d6d6bcc8b1b, 'name': SearchDatastore_Task, 'duration_secs': 0.007762} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.388424] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66f063c9-1e5f-4d2b-a7e2-99aed39757fe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.394038] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a304d0fb-b61c-4c05-a69f-c7307db49ae4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.399456] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 910.399456] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d234e2-e1de-d3f5-1689-f3688fe0d7c1" [ 910.399456] env[61985]: _type = "Task" [ 910.399456] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.404559] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.405025] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 910.405025] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520dca6b-c355-da39-e9e5-20232dc0001f" [ 910.405025] env[61985]: _type = "Task" [ 910.405025] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.416472] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d234e2-e1de-d3f5-1689-f3688fe0d7c1, 'name': SearchDatastore_Task, 'duration_secs': 0.009094} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.421067] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.421067] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3/aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 910.421238] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520dca6b-c355-da39-e9e5-20232dc0001f, 'name': SearchDatastore_Task, 'duration_secs': 0.00841} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.421447] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eda0bdc4-fd10-4666-b70b-d5ddc1842d95 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.424305] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.424428] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 073a7668-39e6-480d-9350-835a0282b456/073a7668-39e6-480d-9350-835a0282b456.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 910.424741] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc76fe22-d348-4b2c-8231-b1aa0d1ca5d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.433272] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 910.433272] env[61985]: value = "task-936054" [ 910.433272] env[61985]: _type = "Task" [ 910.433272] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.433272] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 910.433272] env[61985]: value = "task-936053" [ 910.433272] env[61985]: _type = "Task" [ 910.433272] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.443891] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936054, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.447124] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936053, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.573486] env[61985]: DEBUG nova.compute.manager [req-573eb678-4bac-449c-a710-8672820b260e req-8e693ef0-cd17-42d9-8517-f6306eda7604 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Received event network-vif-deleted-5b4f0fea-8589-4efc-b9b3-b59de5621860 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 910.573707] env[61985]: INFO nova.compute.manager [req-573eb678-4bac-449c-a710-8672820b260e req-8e693ef0-cd17-42d9-8517-f6306eda7604 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Neutron deleted interface 5b4f0fea-8589-4efc-b9b3-b59de5621860; detaching it from the instance and deleting it from the info cache [ 910.573890] env[61985]: DEBUG nova.network.neutron [req-573eb678-4bac-449c-a710-8672820b260e req-8e693ef0-cd17-42d9-8517-f6306eda7604 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.659730] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.755353] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936052, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.904737] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.951984] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936054, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.954857] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936053, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.033991] env[61985]: DEBUG nova.network.neutron [-] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.085363] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9ed388b-6333-49e5-aaae-4c93681609ba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.101167] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ec3fcf-fd72-43c0-b99b-4cb1101acff7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.139405] env[61985]: DEBUG nova.compute.manager [req-573eb678-4bac-449c-a710-8672820b260e req-8e693ef0-cd17-42d9-8517-f6306eda7604 service nova] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Detach interface failed, port_id=5b4f0fea-8589-4efc-b9b3-b59de5621860, reason: Instance e68cc5ff-83b7-4602-be50-ee4b1fa9ac65 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 911.211037] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a836197d-a541-44b9-8536-cd896d4754a4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.217354] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27e72a3-8aa1-403a-8700-d57affc5733c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.251606] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292d142d-c6aa-43cb-8965-b8df3f0be916 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.263555] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709ed34f-c984-4460-92d3-5ce0c391d3fb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.269127] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936052, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.280099] env[61985]: DEBUG nova.compute.provider_tree [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.398029] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.450149] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936053, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.686854} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.453252] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3/aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 911.453484] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.453753] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936054, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.763053} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.453953] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6500952-2b7d-488f-beef-64022039c7ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.455812] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 073a7668-39e6-480d-9350-835a0282b456/073a7668-39e6-480d-9350-835a0282b456.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 911.456033] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.456269] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-42264aa1-c503-4b2e-85a5-76cd6d6c2cf4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.462454] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 911.462454] env[61985]: value = "task-936056" [ 911.462454] env[61985]: _type = "Task" [ 911.462454] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.463633] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 911.463633] env[61985]: value = "task-936057" [ 911.463633] env[61985]: _type = "Task" [ 911.463633] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.473644] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936056, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.476602] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936057, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.537434] env[61985]: INFO nova.compute.manager [-] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Took 1.35 seconds to deallocate network for instance. [ 911.761886] env[61985]: DEBUG oslo_vmware.api [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936052, 'name': RemoveSnapshot_Task, 'duration_secs': 1.943465} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.761990] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 911.762331] env[61985]: INFO nova.compute.manager [None req-b020c50b-b849-4d48-92a3-d89e9f8b1081 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Took 17.17 seconds to snapshot the instance on the hypervisor. [ 911.783307] env[61985]: DEBUG nova.scheduler.client.report [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.900126] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.978630] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936057, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083823} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.979053] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936056, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064446} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.979385] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.979718] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.980579] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba3d8b1-9fce-4764-a55e-108cb6d39b0b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.983344] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d79a7f-559b-4b5d-9a80-657e15e535b1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.014463] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 073a7668-39e6-480d-9350-835a0282b456/073a7668-39e6-480d-9350-835a0282b456.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.024524] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-857eb727-72ac-4841-82a2-c2d26d09a44c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.048512] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3/aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.048867] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f60ebff-307e-4967-a069-66c3d481389b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.063991] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.068225] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 912.068225] env[61985]: value = "task-936058" [ 912.068225] env[61985]: _type = "Task" [ 912.068225] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.069752] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 912.069752] env[61985]: value = "task-936059" [ 912.069752] env[61985]: _type = "Task" [ 912.069752] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.080444] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936059, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.083431] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936058, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.291397] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.597s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.291963] env[61985]: DEBUG nova.compute.manager [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 912.299932] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.379s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.301534] env[61985]: INFO nova.compute.claims [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.406647] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.582601] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936059, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.585634] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936058, 'name': ReconfigVM_Task, 'duration_secs': 0.261059} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.585904] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 073a7668-39e6-480d-9350-835a0282b456/073a7668-39e6-480d-9350-835a0282b456.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.586524] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e067b02c-7081-47ed-a16b-193d0ba6e0a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.593467] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 912.593467] env[61985]: value = "task-936060" [ 912.593467] env[61985]: _type = "Task" [ 912.593467] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.606935] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936060, 'name': Rename_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.801208] env[61985]: DEBUG nova.compute.utils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 912.802679] env[61985]: DEBUG nova.compute.manager [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 912.802933] env[61985]: DEBUG nova.network.neutron [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 912.872533] env[61985]: DEBUG nova.policy [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da7c7d5d10b34b4ebdede5ff3cfd8b01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e66b7a6e7354b06a77295eadf4e5e0f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 912.901196] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.084029] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936059, 'name': ReconfigVM_Task, 'duration_secs': 0.537202} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.084346] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Reconfigured VM instance instance-00000045 to attach disk [datastore1] aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3/aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.084959] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88543868-939c-408f-8466-a5fe0332dc9e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.091438] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 913.091438] env[61985]: value = "task-936061" [ 913.091438] env[61985]: _type = "Task" [ 913.091438] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.103621] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936061, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.106803] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936060, 'name': Rename_Task, 'duration_secs': 0.143289} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.109127] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 913.109127] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d23670f-3d59-441e-8342-30f186d2064e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.115755] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 913.115755] env[61985]: value = "task-936062" [ 913.115755] env[61985]: _type = "Task" [ 913.115755] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.122885] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936062, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.224331] env[61985]: DEBUG nova.network.neutron [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Successfully created port: 334af3b0-2de9-4794-890e-ea3f7b0dd351 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 913.306089] env[61985]: DEBUG nova.compute.manager [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 913.413031] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.603499] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936061, 'name': Rename_Task, 'duration_secs': 0.140089} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.603871] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 913.605502] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e139a8cd-bbe1-4659-a659-609d24b96691 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.614082] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 913.614082] env[61985]: value = "task-936063" [ 913.614082] env[61985]: _type = "Task" [ 913.614082] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.631104] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936063, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.631408] env[61985]: DEBUG oslo_vmware.api [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936062, 'name': PowerOnVM_Task, 'duration_secs': 0.453451} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.633999] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 913.634278] env[61985]: INFO nova.compute.manager [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Took 10.47 seconds to spawn the instance on the hypervisor. [ 913.634550] env[61985]: DEBUG nova.compute.manager [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.635681] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37c0ee1-dd1a-4756-a99e-56a1c1dc85ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.667662] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 913.667662] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211401', 'volume_id': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'name': 'volume-2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1f284789-1e7b-4e9f-9670-34e8e25cd797', 'attached_at': '', 'detached_at': '', 'volume_id': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'serial': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 913.667662] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f0cf00-7f72-4461-bd55-611f83dde534 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.693818] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32df5d0-c56e-41ba-8fe7-47a41115d9dc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.721705] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] volume-2e7a203f-be54-4d0d-b6df-a5b2894cd4d4/volume-2e7a203f-be54-4d0d-b6df-a5b2894cd4d4.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.723501] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d3ac50e-14ca-4319-b0c5-20154800f6c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.738168] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5f4e1a-9715-4bca-9dbf-4c7feedc37bc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.747574] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff499053-c4ec-4421-8a2e-520bb3911a1b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.751815] env[61985]: DEBUG oslo_vmware.api [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 913.751815] env[61985]: value = "task-936064" [ 913.751815] env[61985]: _type = "Task" [ 913.751815] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.783188] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386eb194-1206-46a6-b330-891113bedfa7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.789120] env[61985]: DEBUG oslo_vmware.api [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936064, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.794088] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4451db4f-c0b1-4169-8e69-67ae7fd034e0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.809040] env[61985]: DEBUG nova.compute.provider_tree [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.906450] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.954423] env[61985]: DEBUG nova.compute.manager [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.955351] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e15ea54-2205-4ef7-a885-7d7ad0d73fb6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.124030] env[61985]: DEBUG oslo_vmware.api [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936063, 'name': PowerOnVM_Task, 'duration_secs': 0.49632} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.124354] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 914.124592] env[61985]: INFO nova.compute.manager [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Took 8.43 seconds to spawn the instance on the hypervisor. [ 914.124806] env[61985]: DEBUG nova.compute.manager [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.125660] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99adaf6-6736-479e-96b3-8c32b19ff7af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.154357] env[61985]: INFO nova.compute.manager [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Took 46.68 seconds to build instance. [ 914.261696] env[61985]: DEBUG oslo_vmware.api [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936064, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.311336] env[61985]: DEBUG nova.scheduler.client.report [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.317082] env[61985]: DEBUG nova.compute.manager [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 914.338077] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.338350] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.338581] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.338817] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.338981] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.339153] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.339366] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.339533] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.339704] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.339875] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.340065] env[61985]: DEBUG nova.virt.hardware [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.340906] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b3b9fa-8ce7-4faf-8822-77a13b16a4c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.348686] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1067d826-f259-464d-8d68-838623985341 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.405505] env[61985]: DEBUG oslo_vmware.api [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936037, 'name': ReconfigVM_Task, 'duration_secs': 8.736291} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.405763] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.405976] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Reconfigured VM to detach interface {{(pid=61985) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 914.466785] env[61985]: INFO nova.compute.manager [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] instance snapshotting [ 914.469756] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afca615c-4631-4345-a171-0a4a4bed4590 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.491146] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99499d44-918e-4d82-8617-cd43bf694002 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.644952] env[61985]: INFO nova.compute.manager [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Took 44.92 seconds to build instance. [ 914.656402] env[61985]: DEBUG oslo_concurrency.lockutils [None req-70df06ed-0d04-4005-8c10-c8f3a8b4aff5 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "073a7668-39e6-480d-9350-835a0282b456" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.175s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.746201] env[61985]: DEBUG nova.compute.manager [req-776e61b1-cb20-4d50-8ca2-ea53f6e1d1f9 req-1c845ccf-f886-47ae-be16-41ad7f8caeea service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received event network-vif-deleted-94987e66-83ff-441e-924c-a937a0c6dc27 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 914.747994] env[61985]: INFO nova.compute.manager [req-776e61b1-cb20-4d50-8ca2-ea53f6e1d1f9 req-1c845ccf-f886-47ae-be16-41ad7f8caeea service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Neutron deleted interface 94987e66-83ff-441e-924c-a937a0c6dc27; detaching it from the instance and deleting it from the info cache [ 914.748179] env[61985]: DEBUG nova.network.neutron [req-776e61b1-cb20-4d50-8ca2-ea53f6e1d1f9 req-1c845ccf-f886-47ae-be16-41ad7f8caeea service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updating instance_info_cache with network_info: [{"id": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "address": "fa:16:3e:68:94:dd", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4959d778-77", "ovs_interfaceid": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b464350a-1222-4ab5-9caf-df714dc8f763", "address": "fa:16:3e:2e:99:64", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb464350a-12", "ovs_interfaceid": "b464350a-1222-4ab5-9caf-df714dc8f763", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.764165] env[61985]: DEBUG oslo_vmware.api [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936064, 'name': ReconfigVM_Task, 'duration_secs': 0.784946} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.764491] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Reconfigured VM instance instance-0000003a to attach disk [datastore1] volume-2e7a203f-be54-4d0d-b6df-a5b2894cd4d4/volume-2e7a203f-be54-4d0d-b6df-a5b2894cd4d4.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.770057] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-776c2270-4770-494c-897d-4fa603675c41 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.788326] env[61985]: DEBUG oslo_vmware.api [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 914.788326] env[61985]: value = "task-936065" [ 914.788326] env[61985]: _type = "Task" [ 914.788326] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.797864] env[61985]: DEBUG oslo_vmware.api [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936065, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.816460] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.816863] env[61985]: DEBUG nova.compute.manager [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 914.819763] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.580s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.821234] env[61985]: INFO nova.compute.claims [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.927016] env[61985]: DEBUG nova.compute.manager [req-d07babbb-b752-43fb-9465-0aabdda2100b req-3c6747be-888d-4b57-b33e-60524d77e841 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Received event network-vif-plugged-334af3b0-2de9-4794-890e-ea3f7b0dd351 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 914.927600] env[61985]: DEBUG oslo_concurrency.lockutils [req-d07babbb-b752-43fb-9465-0aabdda2100b req-3c6747be-888d-4b57-b33e-60524d77e841 service nova] Acquiring lock "986c576e-be02-48ac-b24c-72edccab25be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.927838] env[61985]: DEBUG oslo_concurrency.lockutils [req-d07babbb-b752-43fb-9465-0aabdda2100b req-3c6747be-888d-4b57-b33e-60524d77e841 service nova] Lock "986c576e-be02-48ac-b24c-72edccab25be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.928158] env[61985]: DEBUG oslo_concurrency.lockutils [req-d07babbb-b752-43fb-9465-0aabdda2100b req-3c6747be-888d-4b57-b33e-60524d77e841 service nova] Lock "986c576e-be02-48ac-b24c-72edccab25be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.928435] env[61985]: DEBUG nova.compute.manager [req-d07babbb-b752-43fb-9465-0aabdda2100b req-3c6747be-888d-4b57-b33e-60524d77e841 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] No waiting events found dispatching network-vif-plugged-334af3b0-2de9-4794-890e-ea3f7b0dd351 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.928577] env[61985]: WARNING nova.compute.manager [req-d07babbb-b752-43fb-9465-0aabdda2100b req-3c6747be-888d-4b57-b33e-60524d77e841 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Received unexpected event network-vif-plugged-334af3b0-2de9-4794-890e-ea3f7b0dd351 for instance with vm_state building and task_state spawning. [ 915.005305] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 915.005665] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bab4ac05-71b0-4d38-a658-59723e1cb690 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.013984] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 915.013984] env[61985]: value = "task-936066" [ 915.013984] env[61985]: _type = "Task" [ 915.013984] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.023739] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936066, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.152027] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3746e8b-244a-4dd5-902f-2ae150d01bc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.784s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.154869] env[61985]: DEBUG nova.network.neutron [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Successfully updated port: 334af3b0-2de9-4794-890e-ea3f7b0dd351 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.251240] env[61985]: DEBUG oslo_concurrency.lockutils [req-776e61b1-cb20-4d50-8ca2-ea53f6e1d1f9 req-1c845ccf-f886-47ae-be16-41ad7f8caeea service nova] Acquiring lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.251447] env[61985]: DEBUG oslo_concurrency.lockutils [req-776e61b1-cb20-4d50-8ca2-ea53f6e1d1f9 req-1c845ccf-f886-47ae-be16-41ad7f8caeea service nova] Acquired lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.252675] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f217cb77-e67a-4c29-80d0-5cd743ad4212 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.275822] env[61985]: DEBUG oslo_concurrency.lockutils [req-776e61b1-cb20-4d50-8ca2-ea53f6e1d1f9 req-1c845ccf-f886-47ae-be16-41ad7f8caeea service nova] Releasing lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.276294] env[61985]: WARNING nova.compute.manager [req-776e61b1-cb20-4d50-8ca2-ea53f6e1d1f9 req-1c845ccf-f886-47ae-be16-41ad7f8caeea service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Detach interface failed, port_id=94987e66-83ff-441e-924c-a937a0c6dc27, reason: No device with interface-id 94987e66-83ff-441e-924c-a937a0c6dc27 exists on VM: nova.exception.NotFound: No device with interface-id 94987e66-83ff-441e-924c-a937a0c6dc27 exists on VM [ 915.298975] env[61985]: DEBUG oslo_vmware.api [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936065, 'name': ReconfigVM_Task, 'duration_secs': 0.174505} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.299536] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211401', 'volume_id': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'name': 'volume-2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1f284789-1e7b-4e9f-9670-34e8e25cd797', 'attached_at': '', 'detached_at': '', 'volume_id': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'serial': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 915.327715] env[61985]: DEBUG nova.compute.utils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.335145] env[61985]: DEBUG nova.compute.manager [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.335145] env[61985]: DEBUG nova.network.neutron [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 915.407307] env[61985]: DEBUG nova.policy [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '248945e574d34c10a4f2387688673613', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '643e8788ab5746cb9d3259031c275261', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.526634] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936066, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.658626] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "refresh_cache-986c576e-be02-48ac-b24c-72edccab25be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.658626] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "refresh_cache-986c576e-be02-48ac-b24c-72edccab25be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.658626] env[61985]: DEBUG nova.network.neutron [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.679013] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.679013] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.679013] env[61985]: DEBUG nova.network.neutron [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.835152] env[61985]: DEBUG nova.compute.manager [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 915.994980] env[61985]: DEBUG nova.network.neutron [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Successfully created port: 1d016adf-58f0-4fc2-a9d5-4e7232356db4 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.027924] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936066, 'name': CreateSnapshot_Task, 'duration_secs': 0.820442} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.028451] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 916.031542] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2adab6-9861-404f-ab80-1e17c08f1129 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.192152] env[61985]: DEBUG nova.network.neutron [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 916.242649] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22525e5-7228-45a8-8ca0-d181cc3b94e6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.250345] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada328ba-78fe-4d3f-9007-c64e831b99e3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.298058] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1c824d-42c4-42fa-bf40-4c475d45fc01 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.308753] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d589d4ef-c506-4415-8ab9-a68031050ab7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.321629] env[61985]: DEBUG nova.compute.provider_tree [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.348788] env[61985]: DEBUG nova.objects.instance [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lazy-loading 'flavor' on Instance uuid 1f284789-1e7b-4e9f-9670-34e8e25cd797 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.391761] env[61985]: DEBUG nova.network.neutron [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Updating instance_info_cache with network_info: [{"id": "334af3b0-2de9-4794-890e-ea3f7b0dd351", "address": "fa:16:3e:9d:72:df", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap334af3b0-2d", "ovs_interfaceid": "334af3b0-2de9-4794-890e-ea3f7b0dd351", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.552562] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 916.554159] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-960f754a-b72b-4dcc-a9fe-ca9aa3df6c96 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.559284] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.559917] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.570928] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 916.570928] env[61985]: value = "task-936067" [ 916.570928] env[61985]: _type = "Task" [ 916.570928] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.582117] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936067, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.591018] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.591492] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.591492] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.591820] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.592045] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.594523] env[61985]: INFO nova.compute.manager [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Terminating instance [ 916.596278] env[61985]: DEBUG nova.compute.manager [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 916.596486] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 916.597321] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5911376-a9b0-4158-a1ae-bdf756aae259 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.606555] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 916.606965] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb0dc922-e43f-427a-b244-bfd7ebc60734 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.614907] env[61985]: DEBUG oslo_vmware.api [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 916.614907] env[61985]: value = "task-936068" [ 916.614907] env[61985]: _type = "Task" [ 916.614907] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.633022] env[61985]: DEBUG oslo_vmware.api [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936068, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.674169] env[61985]: INFO nova.network.neutron [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Port b464350a-1222-4ab5-9caf-df714dc8f763 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 916.674552] env[61985]: DEBUG nova.network.neutron [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updating instance_info_cache with network_info: [{"id": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "address": "fa:16:3e:68:94:dd", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4959d778-77", "ovs_interfaceid": "4959d778-7709-41bf-a558-f6ba7cce7d4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.826111] env[61985]: DEBUG nova.scheduler.client.report [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.847184] env[61985]: DEBUG nova.compute.manager [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 916.855021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-36508128-2ba1-4459-96f7-6a3283210377 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.331s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.878825] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:29:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='d8c8cc63-01d7-4c63-90a9-d468b4fc7cea',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1410428606',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.879164] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.879352] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.879593] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.880147] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.880147] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.880296] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.880501] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.880732] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.881225] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.881225] env[61985]: DEBUG nova.virt.hardware [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.882408] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4d6140-047b-44eb-be90-874a7154ee3b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.890717] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3bf36f2-5828-4ad5-868b-90d489a99f13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.895169] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "refresh_cache-986c576e-be02-48ac-b24c-72edccab25be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.895413] env[61985]: DEBUG nova.compute.manager [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Instance network_info: |[{"id": "334af3b0-2de9-4794-890e-ea3f7b0dd351", "address": "fa:16:3e:9d:72:df", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap334af3b0-2d", "ovs_interfaceid": "334af3b0-2de9-4794-890e-ea3f7b0dd351", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 916.896665] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:72:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2c019b6-3ef3-4c8f-95bd-edede2c554a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '334af3b0-2de9-4794-890e-ea3f7b0dd351', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 916.903603] env[61985]: DEBUG oslo.service.loopingcall [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.904257] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 916.904517] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b87da4a-10b6-4a27-9bec-fdc6eb3645da {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.933723] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 916.933723] env[61985]: value = "task-936069" [ 916.933723] env[61985]: _type = "Task" [ 916.933723] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.941680] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936069, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.993566] env[61985]: DEBUG nova.compute.manager [req-c05e8586-519a-4f0e-9f00-b0c483683147 req-640fb534-a300-4dc5-a503-72a865774122 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Received event network-changed-a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 916.993795] env[61985]: DEBUG nova.compute.manager [req-c05e8586-519a-4f0e-9f00-b0c483683147 req-640fb534-a300-4dc5-a503-72a865774122 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Refreshing instance network info cache due to event network-changed-a5e0f361-c876-4212-8c4a-dfdfb17d5d4f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 916.994073] env[61985]: DEBUG oslo_concurrency.lockutils [req-c05e8586-519a-4f0e-9f00-b0c483683147 req-640fb534-a300-4dc5-a503-72a865774122 service nova] Acquiring lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.994254] env[61985]: DEBUG oslo_concurrency.lockutils [req-c05e8586-519a-4f0e-9f00-b0c483683147 req-640fb534-a300-4dc5-a503-72a865774122 service nova] Acquired lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.994436] env[61985]: DEBUG nova.network.neutron [req-c05e8586-519a-4f0e-9f00-b0c483683147 req-640fb534-a300-4dc5-a503-72a865774122 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Refreshing network info cache for port a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 917.068054] env[61985]: DEBUG nova.compute.utils [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.081393] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936067, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.124274] env[61985]: DEBUG oslo_vmware.api [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936068, 'name': PowerOffVM_Task, 'duration_secs': 0.223239} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.124546] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 917.124723] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 917.124982] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c452485-065a-44d8-ba1b-b668206f57bc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.156077] env[61985]: DEBUG nova.compute.manager [req-0098167c-50f7-423a-9be5-5e33e8bcf72d req-396be9ff-2d5b-4871-9613-b18aa3b315a5 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Received event network-changed-334af3b0-2de9-4794-890e-ea3f7b0dd351 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 917.156285] env[61985]: DEBUG nova.compute.manager [req-0098167c-50f7-423a-9be5-5e33e8bcf72d req-396be9ff-2d5b-4871-9613-b18aa3b315a5 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Refreshing instance network info cache due to event network-changed-334af3b0-2de9-4794-890e-ea3f7b0dd351. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 917.156493] env[61985]: DEBUG oslo_concurrency.lockutils [req-0098167c-50f7-423a-9be5-5e33e8bcf72d req-396be9ff-2d5b-4871-9613-b18aa3b315a5 service nova] Acquiring lock "refresh_cache-986c576e-be02-48ac-b24c-72edccab25be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.156695] env[61985]: DEBUG oslo_concurrency.lockutils [req-0098167c-50f7-423a-9be5-5e33e8bcf72d req-396be9ff-2d5b-4871-9613-b18aa3b315a5 service nova] Acquired lock "refresh_cache-986c576e-be02-48ac-b24c-72edccab25be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.156799] env[61985]: DEBUG nova.network.neutron [req-0098167c-50f7-423a-9be5-5e33e8bcf72d req-396be9ff-2d5b-4871-9613-b18aa3b315a5 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Refreshing network info cache for port 334af3b0-2de9-4794-890e-ea3f7b0dd351 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 917.178962] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.259658] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 917.260030] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 917.260290] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Deleting the datastore file [datastore1] 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.260599] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1471104c-5ec2-4e23-bab1-edbdea5a3207 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.268142] env[61985]: DEBUG oslo_vmware.api [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 917.268142] env[61985]: value = "task-936071" [ 917.268142] env[61985]: _type = "Task" [ 917.268142] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.277553] env[61985]: DEBUG oslo_vmware.api [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936071, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.332031] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.332614] env[61985]: DEBUG nova.compute.manager [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 917.335919] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.229s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.337045] env[61985]: INFO nova.compute.claims [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.446603] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936069, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.569851] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.581550] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936067, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.683057] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d61fa1c5-9592-402e-8f5b-b2f440ac5db9 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf-94987e66-83ff-441e-924c-a937a0c6dc27" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 12.898s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.781189] env[61985]: DEBUG oslo_vmware.api [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936071, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.348861} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.783249] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.783249] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 917.783249] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 917.783249] env[61985]: INFO nova.compute.manager [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Took 1.19 seconds to destroy the instance on the hypervisor. [ 917.783249] env[61985]: DEBUG oslo.service.loopingcall [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.783249] env[61985]: DEBUG nova.compute.manager [-] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.783249] env[61985]: DEBUG nova.network.neutron [-] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 917.841175] env[61985]: DEBUG nova.compute.utils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.844998] env[61985]: DEBUG nova.compute.manager [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 917.845189] env[61985]: DEBUG nova.network.neutron [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 917.888455] env[61985]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port b464350a-1222-4ab5-9caf-df714dc8f763 could not be found.", "detail": ""}} {{(pid=61985) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 917.888746] env[61985]: DEBUG nova.network.neutron [-] Unable to show port b464350a-1222-4ab5-9caf-df714dc8f763 as it no longer exists. {{(pid=61985) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 917.923195] env[61985]: DEBUG nova.policy [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72dbd4382a5b4656a2ca6768e1b7edc6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be4f1b894ce04f37a0d106e93f4101fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 917.929619] env[61985]: DEBUG nova.network.neutron [req-c05e8586-519a-4f0e-9f00-b0c483683147 req-640fb534-a300-4dc5-a503-72a865774122 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Updated VIF entry in instance network info cache for port a5e0f361-c876-4212-8c4a-dfdfb17d5d4f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 917.929959] env[61985]: DEBUG nova.network.neutron [req-c05e8586-519a-4f0e-9f00-b0c483683147 req-640fb534-a300-4dc5-a503-72a865774122 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Updating instance_info_cache with network_info: [{"id": "a5e0f361-c876-4212-8c4a-dfdfb17d5d4f", "address": "fa:16:3e:3a:9a:10", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5e0f361-c8", "ovs_interfaceid": "a5e0f361-c876-4212-8c4a-dfdfb17d5d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.931781] env[61985]: DEBUG nova.network.neutron [req-0098167c-50f7-423a-9be5-5e33e8bcf72d req-396be9ff-2d5b-4871-9613-b18aa3b315a5 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Updated VIF entry in instance network info cache for port 334af3b0-2de9-4794-890e-ea3f7b0dd351. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 917.932121] env[61985]: DEBUG nova.network.neutron [req-0098167c-50f7-423a-9be5-5e33e8bcf72d req-396be9ff-2d5b-4871-9613-b18aa3b315a5 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Updating instance_info_cache with network_info: [{"id": "334af3b0-2de9-4794-890e-ea3f7b0dd351", "address": "fa:16:3e:9d:72:df", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap334af3b0-2d", "ovs_interfaceid": "334af3b0-2de9-4794-890e-ea3f7b0dd351", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.944923] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936069, 'name': CreateVM_Task, 'duration_secs': 0.574832} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.945825] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 917.946530] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.946700] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.947031] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 917.947346] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a630a8f-1531-4a1b-b3c1-47a3f0695d09 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.952599] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 917.952599] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52250675-9736-10f5-f000-d70e4ff28a0a" [ 917.952599] env[61985]: _type = "Task" [ 917.952599] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.963780] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52250675-9736-10f5-f000-d70e4ff28a0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.086078] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936067, 'name': CloneVM_Task} progress is 95%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.195984] env[61985]: DEBUG nova.network.neutron [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Successfully created port: 47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.201817] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.202296] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.346478] env[61985]: DEBUG nova.compute.manager [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 918.349422] env[61985]: DEBUG nova.compute.manager [req-d86c157f-af17-4e18-a4eb-ea1dc8778a50 req-13e2ae80-ca73-48e5-8cbf-ece8f8abfa07 service nova] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Received event network-vif-plugged-1d016adf-58f0-4fc2-a9d5-4e7232356db4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 918.349637] env[61985]: DEBUG oslo_concurrency.lockutils [req-d86c157f-af17-4e18-a4eb-ea1dc8778a50 req-13e2ae80-ca73-48e5-8cbf-ece8f8abfa07 service nova] Acquiring lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.349850] env[61985]: DEBUG oslo_concurrency.lockutils [req-d86c157f-af17-4e18-a4eb-ea1dc8778a50 req-13e2ae80-ca73-48e5-8cbf-ece8f8abfa07 service nova] Lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.350031] env[61985]: DEBUG oslo_concurrency.lockutils [req-d86c157f-af17-4e18-a4eb-ea1dc8778a50 req-13e2ae80-ca73-48e5-8cbf-ece8f8abfa07 service nova] Lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.350204] env[61985]: DEBUG nova.compute.manager [req-d86c157f-af17-4e18-a4eb-ea1dc8778a50 req-13e2ae80-ca73-48e5-8cbf-ece8f8abfa07 service nova] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] No waiting events found dispatching network-vif-plugged-1d016adf-58f0-4fc2-a9d5-4e7232356db4 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.350372] env[61985]: WARNING nova.compute.manager [req-d86c157f-af17-4e18-a4eb-ea1dc8778a50 req-13e2ae80-ca73-48e5-8cbf-ece8f8abfa07 service nova] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Received unexpected event network-vif-plugged-1d016adf-58f0-4fc2-a9d5-4e7232356db4 for instance with vm_state building and task_state spawning. [ 918.436362] env[61985]: DEBUG oslo_concurrency.lockutils [req-c05e8586-519a-4f0e-9f00-b0c483683147 req-640fb534-a300-4dc5-a503-72a865774122 service nova] Releasing lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.436651] env[61985]: DEBUG nova.compute.manager [req-c05e8586-519a-4f0e-9f00-b0c483683147 req-640fb534-a300-4dc5-a503-72a865774122 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received event network-vif-deleted-b464350a-1222-4ab5-9caf-df714dc8f763 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 918.437216] env[61985]: DEBUG oslo_concurrency.lockutils [req-0098167c-50f7-423a-9be5-5e33e8bcf72d req-396be9ff-2d5b-4871-9613-b18aa3b315a5 service nova] Releasing lock "refresh_cache-986c576e-be02-48ac-b24c-72edccab25be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.468840] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52250675-9736-10f5-f000-d70e4ff28a0a, 'name': SearchDatastore_Task, 'duration_secs': 0.014201} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.469145] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.469376] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.469611] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.469790] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.469932] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.470581] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a697e04-09af-49af-a7d3-f5128803e916 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.481917] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.482155] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 918.483198] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4e9f7c0-d8d5-4973-8b38-3554b792ad26 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.493031] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 918.493031] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52165edb-4e31-5006-0042-4a6d6b97a057" [ 918.493031] env[61985]: _type = "Task" [ 918.493031] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.502201] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52165edb-4e31-5006-0042-4a6d6b97a057, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.531879] env[61985]: DEBUG nova.network.neutron [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Successfully updated port: 1d016adf-58f0-4fc2-a9d5-4e7232356db4 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.586318] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936067, 'name': CloneVM_Task, 'duration_secs': 1.655399} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.586592] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Created linked-clone VM from snapshot [ 918.587890] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471f38a4-b619-4b44-951c-dffa48d687b7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.597457] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Uploading image 71734001-1b38-49c6-84ab-bdbcf8ca062f {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 918.629468] env[61985]: DEBUG oslo_vmware.rw_handles [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 918.629468] env[61985]: value = "vm-211404" [ 918.629468] env[61985]: _type = "VirtualMachine" [ 918.629468] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 918.629763] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b8580b57-325b-4839-b1ce-d99cdc387cfa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.631616] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.631847] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.632084] env[61985]: INFO nova.compute.manager [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Attaching volume 6c7f9ac2-d6bd-479b-86cb-5641a612d6bb to /dev/sdb [ 918.641928] env[61985]: DEBUG oslo_vmware.rw_handles [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lease: (returnval){ [ 918.641928] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524ba202-becd-2439-c101-0112de79c27e" [ 918.641928] env[61985]: _type = "HttpNfcLease" [ 918.641928] env[61985]: } obtained for exporting VM: (result){ [ 918.641928] env[61985]: value = "vm-211404" [ 918.641928] env[61985]: _type = "VirtualMachine" [ 918.641928] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 918.641928] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the lease: (returnval){ [ 918.641928] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524ba202-becd-2439-c101-0112de79c27e" [ 918.641928] env[61985]: _type = "HttpNfcLease" [ 918.641928] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 918.648848] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 918.648848] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524ba202-becd-2439-c101-0112de79c27e" [ 918.648848] env[61985]: _type = "HttpNfcLease" [ 918.648848] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 918.678225] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973af5d1-8b4a-4ed2-bd16-a362c97f8b5d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.689911] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bcde21-004d-43e9-8a1e-d5259e15c3d8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.703497] env[61985]: DEBUG nova.virt.block_device [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Updating existing volume attachment record: 55ec25cb-765c-48ed-9987-9e777e9658ed {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 918.707149] env[61985]: DEBUG nova.compute.utils [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.772581] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05abd877-4ac0-45e7-8503-ca835be94eff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.779641] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6959a3f3-909b-4305-a08f-e45f73ccf7b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.812648] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7447e0a6-8c2d-484c-9c3c-cead08331a56 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.821573] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deebc461-8bd9-4ebd-86fb-b6ad6a71112f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.835266] env[61985]: DEBUG nova.compute.provider_tree [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.003114] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52165edb-4e31-5006-0042-4a6d6b97a057, 'name': SearchDatastore_Task, 'duration_secs': 0.008541} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.003799] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-331763d4-c36b-4532-a7f2-665c5221a225 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.008706] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 919.008706] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526be4c9-04f2-1ac4-b381-1dc51a32c29b" [ 919.008706] env[61985]: _type = "Task" [ 919.008706] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.016542] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526be4c9-04f2-1ac4-b381-1dc51a32c29b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.034889] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.034984] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.035158] env[61985]: DEBUG nova.network.neutron [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 919.080857] env[61985]: DEBUG nova.network.neutron [-] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.150497] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 919.150497] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524ba202-becd-2439-c101-0112de79c27e" [ 919.150497] env[61985]: _type = "HttpNfcLease" [ 919.150497] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 919.150812] env[61985]: DEBUG oslo_vmware.rw_handles [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 919.150812] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524ba202-becd-2439-c101-0112de79c27e" [ 919.150812] env[61985]: _type = "HttpNfcLease" [ 919.150812] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 919.151572] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9664e74-16e9-4e65-81d6-feb6056671b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.159216] env[61985]: DEBUG oslo_vmware.rw_handles [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d818f8-8f6f-9333-ace3-de745a5aaf1e/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 919.159415] env[61985]: DEBUG oslo_vmware.rw_handles [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d818f8-8f6f-9333-ace3-de745a5aaf1e/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 919.220375] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.018s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.224982] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.225221] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.254603] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b5e7eae7-16bd-401a-bfed-0ef0d7da19ce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.338304] env[61985]: DEBUG nova.scheduler.client.report [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.357252] env[61985]: DEBUG nova.compute.manager [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 919.378518] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.378794] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.378965] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.379169] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.379323] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.379476] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.379688] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.379933] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.380037] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.380214] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.380393] env[61985]: DEBUG nova.virt.hardware [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.382431] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4738a7ce-4633-4e90-b66e-6981f60ebecc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.390751] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733d1530-bdda-42df-b2e4-21317015cbde {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.227872] env[61985]: INFO nova.compute.manager [-] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Took 2.45 seconds to deallocate network for instance. [ 920.229387] env[61985]: DEBUG nova.compute.utils [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.232819] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.898s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.233347] env[61985]: DEBUG nova.compute.manager [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 920.239619] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526be4c9-04f2-1ac4-b381-1dc51a32c29b, 'name': SearchDatastore_Task, 'duration_secs': 0.008944} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.239619] env[61985]: DEBUG nova.network.neutron [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Successfully updated port: 47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 920.242305] env[61985]: DEBUG nova.compute.manager [req-e3056a43-d972-4225-8faf-e570b0c3db2a req-23cf8222-0d98-44e2-b89d-088328aca157 service nova] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Received event network-changed-1d016adf-58f0-4fc2-a9d5-4e7232356db4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 920.242566] env[61985]: DEBUG nova.compute.manager [req-e3056a43-d972-4225-8faf-e570b0c3db2a req-23cf8222-0d98-44e2-b89d-088328aca157 service nova] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Refreshing instance network info cache due to event network-changed-1d016adf-58f0-4fc2-a9d5-4e7232356db4. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 920.242798] env[61985]: DEBUG oslo_concurrency.lockutils [req-e3056a43-d972-4225-8faf-e570b0c3db2a req-23cf8222-0d98-44e2-b89d-088328aca157 service nova] Acquiring lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.244102] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.649s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.244348] env[61985]: INFO nova.compute.claims [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.248182] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.248426] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 986c576e-be02-48ac-b24c-72edccab25be/986c576e-be02-48ac-b24c-72edccab25be.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 920.252711] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b217440e-6af3-49b1-858e-a939dfba619c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.262599] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 920.262599] env[61985]: value = "task-936076" [ 920.262599] env[61985]: _type = "Task" [ 920.262599] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.273506] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936076, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.301801] env[61985]: DEBUG nova.network.neutron [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 920.450290] env[61985]: DEBUG nova.compute.manager [req-54f721f9-fe89-4bf2-988b-560fd6d52b42 req-24f9f5bb-c068-4c83-a55f-2162e87d4bee service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Received event network-vif-plugged-47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 920.450665] env[61985]: DEBUG oslo_concurrency.lockutils [req-54f721f9-fe89-4bf2-988b-560fd6d52b42 req-24f9f5bb-c068-4c83-a55f-2162e87d4bee service nova] Acquiring lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.450906] env[61985]: DEBUG oslo_concurrency.lockutils [req-54f721f9-fe89-4bf2-988b-560fd6d52b42 req-24f9f5bb-c068-4c83-a55f-2162e87d4bee service nova] Lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.451104] env[61985]: DEBUG oslo_concurrency.lockutils [req-54f721f9-fe89-4bf2-988b-560fd6d52b42 req-24f9f5bb-c068-4c83-a55f-2162e87d4bee service nova] Lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.451283] env[61985]: DEBUG nova.compute.manager [req-54f721f9-fe89-4bf2-988b-560fd6d52b42 req-24f9f5bb-c068-4c83-a55f-2162e87d4bee service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] No waiting events found dispatching network-vif-plugged-47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.451454] env[61985]: WARNING nova.compute.manager [req-54f721f9-fe89-4bf2-988b-560fd6d52b42 req-24f9f5bb-c068-4c83-a55f-2162e87d4bee service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Received unexpected event network-vif-plugged-47c6b710-9517-4551-b70b-d11e6cdd7c8f for instance with vm_state building and task_state spawning. [ 920.532896] env[61985]: DEBUG nova.network.neutron [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance_info_cache with network_info: [{"id": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "address": "fa:16:3e:b5:b4:b7", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.147", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d016adf-58", "ovs_interfaceid": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.738147] env[61985]: DEBUG nova.compute.utils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.740379] env[61985]: DEBUG nova.compute.manager [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 920.740566] env[61985]: DEBUG nova.network.neutron [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 920.756382] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.756829] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.757164] env[61985]: INFO nova.compute.manager [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Attaching volume 78b9d30c-3370-4505-b9bf-60bae00d9eda to /dev/sdc [ 920.760131] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.763386] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.763564] env[61985]: DEBUG nova.network.neutron [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 920.765348] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.540s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.781781] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.791913] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936076, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.821715] env[61985]: DEBUG nova.policy [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e23fdb33638143409613d9d4f26f199c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe531ec9f60548adb457351e19f8d51b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.825718] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4df6e6e-6eeb-4856-af35-929b20028a02 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.833973] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9f59bc-3660-4088-ba15-349ca448e463 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.848123] env[61985]: DEBUG nova.virt.block_device [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Updating existing volume attachment record: 75f55c26-97cb-4ecb-ba17-efde8067df79 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 921.035431] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.035986] env[61985]: DEBUG nova.compute.manager [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Instance network_info: |[{"id": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "address": "fa:16:3e:b5:b4:b7", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.147", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d016adf-58", "ovs_interfaceid": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 921.036366] env[61985]: DEBUG oslo_concurrency.lockutils [req-e3056a43-d972-4225-8faf-e570b0c3db2a req-23cf8222-0d98-44e2-b89d-088328aca157 service nova] Acquired lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.036552] env[61985]: DEBUG nova.network.neutron [req-e3056a43-d972-4225-8faf-e570b0c3db2a req-23cf8222-0d98-44e2-b89d-088328aca157 service nova] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Refreshing network info cache for port 1d016adf-58f0-4fc2-a9d5-4e7232356db4 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 921.042536] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:b4:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e1049e8-c06b-4c93-a9e1-2cbb530f3f95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d016adf-58f0-4fc2-a9d5-4e7232356db4', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.056969] env[61985]: DEBUG oslo.service.loopingcall [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.058178] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 921.060762] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f5c6ecd-1ba9-4739-83cf-a796e784c785 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.094433] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.094433] env[61985]: value = "task-936078" [ 921.094433] env[61985]: _type = "Task" [ 921.094433] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.100365] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936078, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.250283] env[61985]: DEBUG nova.compute.manager [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 921.269617] env[61985]: DEBUG nova.network.neutron [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Successfully created port: 41b487bd-afda-4a0c-be1a-056e516b010c {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.284839] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936076, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.61641} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.288087] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 986c576e-be02-48ac-b24c-72edccab25be/986c576e-be02-48ac-b24c-72edccab25be.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 921.288413] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 921.289107] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2fc37895-53b5-47a5-bd51-fa1737835c18 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.296512] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 921.296512] env[61985]: value = "task-936080" [ 921.296512] env[61985]: _type = "Task" [ 921.296512] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.306028] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936080, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.314372] env[61985]: DEBUG nova.network.neutron [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 921.484388] env[61985]: DEBUG nova.network.neutron [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updating instance_info_cache with network_info: [{"id": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "address": "fa:16:3e:af:c8:fd", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47c6b710-95", "ovs_interfaceid": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.607310] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936078, 'name': CreateVM_Task, 'duration_secs': 0.31409} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.610601] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 921.611866] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.612193] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.612646] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.613142] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29a4b37e-7bb5-40b5-8423-2a96b38efe0f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.618248] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 921.618248] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521f964d-6869-41f0-aea6-aafaae701526" [ 921.618248] env[61985]: _type = "Task" [ 921.618248] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.629558] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521f964d-6869-41f0-aea6-aafaae701526, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.699610] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61005dc9-088d-4e58-99c9-a01c885a5fdb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.708189] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b4cce1-97b2-48b7-a639-0925f4d2862d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.746666] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a627f89-24a7-4eb9-a627-14182c7e38a2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.756297] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d3a33e-e88f-4a12-9f26-bbc7440d17e0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.776505] env[61985]: DEBUG nova.compute.provider_tree [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.808555] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936080, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070535} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.809161] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.809757] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba6d8bf-c253-4e6f-bf61-9e1f7ddf60d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.835989] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 986c576e-be02-48ac-b24c-72edccab25be/986c576e-be02-48ac-b24c-72edccab25be.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.836742] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2c4ca4e-e802-43d5-9206-eb644428c2ed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.853007] env[61985]: DEBUG nova.network.neutron [req-e3056a43-d972-4225-8faf-e570b0c3db2a req-23cf8222-0d98-44e2-b89d-088328aca157 service nova] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updated VIF entry in instance network info cache for port 1d016adf-58f0-4fc2-a9d5-4e7232356db4. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 921.853480] env[61985]: DEBUG nova.network.neutron [req-e3056a43-d972-4225-8faf-e570b0c3db2a req-23cf8222-0d98-44e2-b89d-088328aca157 service nova] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance_info_cache with network_info: [{"id": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "address": "fa:16:3e:b5:b4:b7", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.147", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d016adf-58", "ovs_interfaceid": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.861056] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 921.861056] env[61985]: value = "task-936081" [ 921.861056] env[61985]: _type = "Task" [ 921.861056] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.871987] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936081, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.987761] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.988145] env[61985]: DEBUG nova.compute.manager [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Instance network_info: |[{"id": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "address": "fa:16:3e:af:c8:fd", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47c6b710-95", "ovs_interfaceid": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 921.988651] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:c8:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cb0e556a-0f69-4a5c-af62-ffc46edb8e63', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47c6b710-9517-4551-b70b-d11e6cdd7c8f', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.996434] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Creating folder: Project (be4f1b894ce04f37a0d106e93f4101fa). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 921.996772] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04aa7cc0-b961-4c08-b627-98577cbf3cbe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.008461] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Created folder: Project (be4f1b894ce04f37a0d106e93f4101fa) in parent group-v211285. [ 922.008806] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Creating folder: Instances. Parent ref: group-v211410. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 922.009196] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8881df0c-fa8b-4fe5-865f-4d68bea17c92 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.019383] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Created folder: Instances in parent group-v211410. [ 922.023018] env[61985]: DEBUG oslo.service.loopingcall [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.023018] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 922.023018] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2859086-818b-4908-a5f4-5e024a2e75b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.044074] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.044074] env[61985]: value = "task-936084" [ 922.044074] env[61985]: _type = "Task" [ 922.044074] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.054295] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936084, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.129898] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521f964d-6869-41f0-aea6-aafaae701526, 'name': SearchDatastore_Task, 'duration_secs': 0.024954} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.130268] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.130673] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.131120] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.131326] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.131568] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.131892] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17f68f85-ced8-448b-b524-2887c4ab4c9d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.140586] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.140812] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 922.141567] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a723a02-d795-4172-92bd-d3b09773d42f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.147363] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 922.147363] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529827ec-15eb-e8fa-36f0-0cee0077196f" [ 922.147363] env[61985]: _type = "Task" [ 922.147363] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.155908] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529827ec-15eb-e8fa-36f0-0cee0077196f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.268801] env[61985]: DEBUG nova.compute.manager [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 922.282267] env[61985]: DEBUG nova.scheduler.client.report [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.295211] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.295523] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.295728] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.295987] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.296160] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.296324] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.297014] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.297014] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.297014] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.297121] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.297311] env[61985]: DEBUG nova.virt.hardware [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.299514] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a139fcf0-51ee-4525-b298-770957282a83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.309735] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe16f18e-8360-4b32-b767-8a02e6fcb554 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.356338] env[61985]: DEBUG oslo_concurrency.lockutils [req-e3056a43-d972-4225-8faf-e570b0c3db2a req-23cf8222-0d98-44e2-b89d-088328aca157 service nova] Releasing lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.356338] env[61985]: DEBUG nova.compute.manager [req-e3056a43-d972-4225-8faf-e570b0c3db2a req-23cf8222-0d98-44e2-b89d-088328aca157 service nova] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Received event network-vif-deleted-4959d778-7709-41bf-a558-f6ba7cce7d4b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 922.371068] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936081, 'name': ReconfigVM_Task, 'duration_secs': 0.394685} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.371371] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 986c576e-be02-48ac-b24c-72edccab25be/986c576e-be02-48ac-b24c-72edccab25be.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.372118] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8718b631-9b13-4d71-9162-012f66203894 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.378120] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 922.378120] env[61985]: value = "task-936085" [ 922.378120] env[61985]: _type = "Task" [ 922.378120] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.386229] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936085, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.555046] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936084, 'name': CreateVM_Task, 'duration_secs': 0.301347} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.555298] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 922.555915] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.556135] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.556465] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.556734] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-323db3a5-e070-4dc5-b099-235298346697 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.561635] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 922.561635] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523df6e4-9cc0-3b5b-dec3-ab4ae4936c0c" [ 922.561635] env[61985]: _type = "Task" [ 922.561635] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.570348] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523df6e4-9cc0-3b5b-dec3-ab4ae4936c0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.596517] env[61985]: DEBUG nova.compute.manager [req-60920dbd-862e-4c99-b183-7f702ee9cec5 req-91f7a581-70f5-4ad1-99cc-845b8e61493b service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Received event network-changed-47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 922.596705] env[61985]: DEBUG nova.compute.manager [req-60920dbd-862e-4c99-b183-7f702ee9cec5 req-91f7a581-70f5-4ad1-99cc-845b8e61493b service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Refreshing instance network info cache due to event network-changed-47c6b710-9517-4551-b70b-d11e6cdd7c8f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 922.596924] env[61985]: DEBUG oslo_concurrency.lockutils [req-60920dbd-862e-4c99-b183-7f702ee9cec5 req-91f7a581-70f5-4ad1-99cc-845b8e61493b service nova] Acquiring lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.597090] env[61985]: DEBUG oslo_concurrency.lockutils [req-60920dbd-862e-4c99-b183-7f702ee9cec5 req-91f7a581-70f5-4ad1-99cc-845b8e61493b service nova] Acquired lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.597258] env[61985]: DEBUG nova.network.neutron [req-60920dbd-862e-4c99-b183-7f702ee9cec5 req-91f7a581-70f5-4ad1-99cc-845b8e61493b service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Refreshing network info cache for port 47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 922.658524] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529827ec-15eb-e8fa-36f0-0cee0077196f, 'name': SearchDatastore_Task, 'duration_secs': 0.010249} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.659482] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5dc15c4d-3b47-4081-9d28-fecbdf9d5262 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.665934] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 922.665934] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524f1295-ff30-f02f-e855-ccbd1d8a9455" [ 922.665934] env[61985]: _type = "Task" [ 922.665934] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.674800] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524f1295-ff30-f02f-e855-ccbd1d8a9455, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.789367] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.789924] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 922.792958] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.149s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.794144] env[61985]: INFO nova.compute.claims [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 922.887863] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936085, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.079233] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523df6e4-9cc0-3b5b-dec3-ab4ae4936c0c, 'name': SearchDatastore_Task, 'duration_secs': 0.013539} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.079547] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.079795] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.080077] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.181069] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524f1295-ff30-f02f-e855-ccbd1d8a9455, 'name': SearchDatastore_Task, 'duration_secs': 0.01325} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.181595] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.181866] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] a77ca23f-b2c0-4822-8e48-3e47e0dadb27/a77ca23f-b2c0-4822-8e48-3e47e0dadb27.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 923.182178] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.182368] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.182608] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f089cb51-7bd1-477a-9594-3294afdb9159 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.185187] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a2982c7-ef35-4d8e-906b-4a54b130cfbb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.192164] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 923.192164] env[61985]: value = "task-936087" [ 923.192164] env[61985]: _type = "Task" [ 923.192164] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.195713] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.195915] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 923.201024] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-069fc596-621f-4e8e-9479-e8cff3107ece {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.202619] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936087, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.207819] env[61985]: DEBUG nova.compute.manager [req-4f4b6a44-fdee-4a9d-8685-adb30c798277 req-f6ef3b3f-fbf1-4c24-ab4e-cd845910f2d3 service nova] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Received event network-vif-plugged-41b487bd-afda-4a0c-be1a-056e516b010c {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 923.208049] env[61985]: DEBUG oslo_concurrency.lockutils [req-4f4b6a44-fdee-4a9d-8685-adb30c798277 req-f6ef3b3f-fbf1-4c24-ab4e-cd845910f2d3 service nova] Acquiring lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.208264] env[61985]: DEBUG oslo_concurrency.lockutils [req-4f4b6a44-fdee-4a9d-8685-adb30c798277 req-f6ef3b3f-fbf1-4c24-ab4e-cd845910f2d3 service nova] Lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.208530] env[61985]: DEBUG oslo_concurrency.lockutils [req-4f4b6a44-fdee-4a9d-8685-adb30c798277 req-f6ef3b3f-fbf1-4c24-ab4e-cd845910f2d3 service nova] Lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.208640] env[61985]: DEBUG nova.compute.manager [req-4f4b6a44-fdee-4a9d-8685-adb30c798277 req-f6ef3b3f-fbf1-4c24-ab4e-cd845910f2d3 service nova] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] No waiting events found dispatching network-vif-plugged-41b487bd-afda-4a0c-be1a-056e516b010c {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.210373] env[61985]: WARNING nova.compute.manager [req-4f4b6a44-fdee-4a9d-8685-adb30c798277 req-f6ef3b3f-fbf1-4c24-ab4e-cd845910f2d3 service nova] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Received unexpected event network-vif-plugged-41b487bd-afda-4a0c-be1a-056e516b010c for instance with vm_state building and task_state spawning. [ 923.210916] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 923.210916] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5220e526-d539-9712-268c-0f432ef8a211" [ 923.210916] env[61985]: _type = "Task" [ 923.210916] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.221551] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5220e526-d539-9712-268c-0f432ef8a211, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.222066] env[61985]: DEBUG nova.network.neutron [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Successfully updated port: 41b487bd-afda-4a0c-be1a-056e516b010c {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.265434] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 923.265916] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211407', 'volume_id': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'name': 'volume-6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'serial': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 923.266569] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce0176c-f0e4-4e6c-a70c-3f76033b6c46 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.288027] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48cc178d-6454-42ad-97d2-d352f26bf968 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.307534] env[61985]: DEBUG nova.compute.utils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 923.319076] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] volume-6c7f9ac2-d6bd-479b-86cb-5641a612d6bb/volume-6c7f9ac2-d6bd-479b-86cb-5641a612d6bb.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.319331] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 923.319640] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 923.321439] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-989e325d-fe7c-4030-9175-6885e6abf0ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.348826] env[61985]: DEBUG oslo_vmware.api [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 923.348826] env[61985]: value = "task-936088" [ 923.348826] env[61985]: _type = "Task" [ 923.348826] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.359658] env[61985]: DEBUG oslo_vmware.api [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936088, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.382525] env[61985]: DEBUG nova.network.neutron [req-60920dbd-862e-4c99-b183-7f702ee9cec5 req-91f7a581-70f5-4ad1-99cc-845b8e61493b service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updated VIF entry in instance network info cache for port 47c6b710-9517-4551-b70b-d11e6cdd7c8f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.382525] env[61985]: DEBUG nova.network.neutron [req-60920dbd-862e-4c99-b183-7f702ee9cec5 req-91f7a581-70f5-4ad1-99cc-845b8e61493b service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updating instance_info_cache with network_info: [{"id": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "address": "fa:16:3e:af:c8:fd", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47c6b710-95", "ovs_interfaceid": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.385394] env[61985]: DEBUG nova.policy [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3d8fab40f554eb3a4804a4291686146', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c06fc8ffa004e00be0a1991d2e082e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 923.399531] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936085, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.710254] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936087, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.731228] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "refresh_cache-5c2a7072-5b14-4ec7-8060-4e21eea927e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.731302] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "refresh_cache-5c2a7072-5b14-4ec7-8060-4e21eea927e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.731548] env[61985]: DEBUG nova.network.neutron [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 923.733376] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5220e526-d539-9712-268c-0f432ef8a211, 'name': SearchDatastore_Task, 'duration_secs': 0.021411} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.735572] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa33e004-34cc-4803-9288-de8b055bf948 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.744216] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 923.744216] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524eafcf-c063-ed95-52f7-b5e03a07e09f" [ 923.744216] env[61985]: _type = "Task" [ 923.744216] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.756806] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524eafcf-c063-ed95-52f7-b5e03a07e09f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.820588] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 923.863154] env[61985]: DEBUG oslo_vmware.api [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936088, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.884527] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.886019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.886019] env[61985]: INFO nova.compute.manager [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Attaching volume 774d6c9b-3a7e-48dc-836f-82cec83d6e53 to /dev/sdb [ 923.891339] env[61985]: DEBUG oslo_concurrency.lockutils [req-60920dbd-862e-4c99-b183-7f702ee9cec5 req-91f7a581-70f5-4ad1-99cc-845b8e61493b service nova] Releasing lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.902606] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936085, 'name': Rename_Task, 'duration_secs': 1.275579} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.902914] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 923.903261] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9340883-7371-4444-b88d-2d8771635eae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.916420] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 923.916420] env[61985]: value = "task-936089" [ 923.916420] env[61985]: _type = "Task" [ 923.916420] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.926138] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936089, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.931174] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8697ff2-3dd2-4665-a934-34dbcb8486da {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.941404] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ed8b01-b073-485b-b3e1-35435cdb6723 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.958066] env[61985]: DEBUG nova.virt.block_device [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Updating existing volume attachment record: 282116ce-675b-4f37-b897-67e1cb7a0f09 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 924.021068] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Successfully created port: 6ee77311-001c-4c2f-afad-e75094a76022 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 924.202380] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936087, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.859766} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.205237] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] a77ca23f-b2c0-4822-8e48-3e47e0dadb27/a77ca23f-b2c0-4822-8e48-3e47e0dadb27.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 924.205912] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 924.206439] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b36db03-8767-4cc5-8734-43a8de087bb3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.212797] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 924.212797] env[61985]: value = "task-936093" [ 924.212797] env[61985]: _type = "Task" [ 924.212797] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.223947] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936093, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.258979] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524eafcf-c063-ed95-52f7-b5e03a07e09f, 'name': SearchDatastore_Task, 'duration_secs': 0.053906} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.259188] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.259464] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 87006fbf-ea90-4d9a-88af-001de424ac14/87006fbf-ea90-4d9a-88af-001de424ac14.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 924.259759] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-821ef7eb-9a6e-46ad-b791-468a3e884c45 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.266935] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 924.266935] env[61985]: value = "task-936094" [ 924.266935] env[61985]: _type = "Task" [ 924.266935] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.271393] env[61985]: DEBUG nova.network.neutron [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 924.281895] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.283634] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64da7e38-aa49-4fbc-a2c7-594affbacb40 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.290584] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc487bf-e8c5-477d-8a30-da369e1bc3cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.326189] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8df9fa8-ec7a-469a-8ff4-e3500ef692b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.342378] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea400c45-143d-4590-aeaa-b34868f3d895 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.357295] env[61985]: DEBUG nova.compute.provider_tree [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.367391] env[61985]: DEBUG oslo_vmware.api [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936088, 'name': ReconfigVM_Task, 'duration_secs': 0.581759} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.367783] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Reconfigured VM instance instance-00000045 to attach disk [datastore1] volume-6c7f9ac2-d6bd-479b-86cb-5641a612d6bb/volume-6c7f9ac2-d6bd-479b-86cb-5641a612d6bb.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.376022] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ab255b6-2c8d-450b-818d-548b80acc17a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.393787] env[61985]: DEBUG oslo_vmware.api [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 924.393787] env[61985]: value = "task-936095" [ 924.393787] env[61985]: _type = "Task" [ 924.393787] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.403484] env[61985]: DEBUG oslo_vmware.api [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936095, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.431770] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936089, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.482640] env[61985]: DEBUG nova.network.neutron [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Updating instance_info_cache with network_info: [{"id": "41b487bd-afda-4a0c-be1a-056e516b010c", "address": "fa:16:3e:a0:47:19", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41b487bd-af", "ovs_interfaceid": "41b487bd-afda-4a0c-be1a-056e516b010c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.726567] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936093, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061257} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.727206] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 924.728428] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8124ee65-4395-451c-9aac-cc1f2113a4df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.753810] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] a77ca23f-b2c0-4822-8e48-3e47e0dadb27/a77ca23f-b2c0-4822-8e48-3e47e0dadb27.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.754136] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5431f6b3-10e7-46bb-a9a9-b66e5d07c4b1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.775688] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 924.775688] env[61985]: value = "task-936096" [ 924.775688] env[61985]: _type = "Task" [ 924.775688] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.778734] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936094, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.787713] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936096, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.842280] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 924.864080] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.864404] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.864528] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.864718] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.865477] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.865477] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.865477] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.865477] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.865707] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.865763] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.865939] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.866874] env[61985]: DEBUG nova.scheduler.client.report [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.870808] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e79a0f9-ea85-43f7-a377-4a8ebcd28a94 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.882045] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b063f19a-767c-452e-8aa3-9fb0bfb901f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.906098] env[61985]: DEBUG oslo_vmware.api [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936095, 'name': ReconfigVM_Task, 'duration_secs': 0.160762} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.906417] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211407', 'volume_id': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'name': 'volume-6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'serial': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 924.927755] env[61985]: DEBUG oslo_vmware.api [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936089, 'name': PowerOnVM_Task, 'duration_secs': 0.543775} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.928061] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 924.928287] env[61985]: INFO nova.compute.manager [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Took 10.61 seconds to spawn the instance on the hypervisor. [ 924.928474] env[61985]: DEBUG nova.compute.manager [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.929257] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8abfd13-92d5-4308-be47-4da2187198e6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.985299] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "refresh_cache-5c2a7072-5b14-4ec7-8060-4e21eea927e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.985697] env[61985]: DEBUG nova.compute.manager [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Instance network_info: |[{"id": "41b487bd-afda-4a0c-be1a-056e516b010c", "address": "fa:16:3e:a0:47:19", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41b487bd-af", "ovs_interfaceid": "41b487bd-afda-4a0c-be1a-056e516b010c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 924.986179] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:47:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55520f67-d092-4eb7-940f-d7cceaa1ca1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41b487bd-afda-4a0c-be1a-056e516b010c', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.993900] env[61985]: DEBUG oslo.service.loopingcall [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.994431] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 924.994675] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c40fac70-2c8d-49e2-9de9-3bb4cd2bd11b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.014681] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.014681] env[61985]: value = "task-936097" [ 925.014681] env[61985]: _type = "Task" [ 925.014681] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.023074] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936097, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.238461] env[61985]: DEBUG nova.compute.manager [req-0bb2fe08-abf3-46ed-92ef-14e2be85ad75 req-c372b531-a398-4487-838a-502a861842d3 service nova] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Received event network-changed-41b487bd-afda-4a0c-be1a-056e516b010c {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 925.238713] env[61985]: DEBUG nova.compute.manager [req-0bb2fe08-abf3-46ed-92ef-14e2be85ad75 req-c372b531-a398-4487-838a-502a861842d3 service nova] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Refreshing instance network info cache due to event network-changed-41b487bd-afda-4a0c-be1a-056e516b010c. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 925.238908] env[61985]: DEBUG oslo_concurrency.lockutils [req-0bb2fe08-abf3-46ed-92ef-14e2be85ad75 req-c372b531-a398-4487-838a-502a861842d3 service nova] Acquiring lock "refresh_cache-5c2a7072-5b14-4ec7-8060-4e21eea927e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.239310] env[61985]: DEBUG oslo_concurrency.lockutils [req-0bb2fe08-abf3-46ed-92ef-14e2be85ad75 req-c372b531-a398-4487-838a-502a861842d3 service nova] Acquired lock "refresh_cache-5c2a7072-5b14-4ec7-8060-4e21eea927e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.239526] env[61985]: DEBUG nova.network.neutron [req-0bb2fe08-abf3-46ed-92ef-14e2be85ad75 req-c372b531-a398-4487-838a-502a861842d3 service nova] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Refreshing network info cache for port 41b487bd-afda-4a0c-be1a-056e516b010c {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 925.281435] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936094, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519791} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.284810] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 87006fbf-ea90-4d9a-88af-001de424ac14/87006fbf-ea90-4d9a-88af-001de424ac14.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.285101] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.285469] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-caad44eb-71a5-4ddc-89c4-b6f2e70dce85 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.292826] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936096, 'name': ReconfigVM_Task, 'duration_secs': 0.324659} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.294101] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Reconfigured VM instance instance-00000047 to attach disk [datastore1] a77ca23f-b2c0-4822-8e48-3e47e0dadb27/a77ca23f-b2c0-4822-8e48-3e47e0dadb27.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.294849] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 925.294849] env[61985]: value = "task-936098" [ 925.294849] env[61985]: _type = "Task" [ 925.294849] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.295097] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e9d4d40-caf4-4b49-9415-f592e8264c72 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.304982] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936098, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.306482] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 925.306482] env[61985]: value = "task-936099" [ 925.306482] env[61985]: _type = "Task" [ 925.306482] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.315410] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936099, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.377671] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.585s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.378262] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 925.381544] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.568s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.381803] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.384309] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.665s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.384504] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.386459] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.645s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.388067] env[61985]: INFO nova.compute.claims [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 925.416693] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 925.417036] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211408', 'volume_id': '78b9d30c-3370-4505-b9bf-60bae00d9eda', 'name': 'volume-78b9d30c-3370-4505-b9bf-60bae00d9eda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1f284789-1e7b-4e9f-9670-34e8e25cd797', 'attached_at': '', 'detached_at': '', 'volume_id': '78b9d30c-3370-4505-b9bf-60bae00d9eda', 'serial': '78b9d30c-3370-4505-b9bf-60bae00d9eda'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 925.418109] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15e1ee9-8c34-48e6-9d2e-3f2a74afaf0f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.423847] env[61985]: INFO nova.scheduler.client.report [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Deleted allocations for instance c64056d8-0c22-48db-bf4a-a60f468bdad0 [ 925.425528] env[61985]: INFO nova.scheduler.client.report [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Deleted allocations for instance 5b3a8ba3-48cb-472a-ab45-4e51af68812c [ 925.449375] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1b27f5-3c64-412d-975a-540975ae54b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.455763] env[61985]: INFO nova.compute.manager [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Took 51.33 seconds to build instance. [ 925.483561] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] volume-78b9d30c-3370-4505-b9bf-60bae00d9eda/volume-78b9d30c-3370-4505-b9bf-60bae00d9eda.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.484909] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6617bf48-043e-4869-a5fb-20ce31878224 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.503708] env[61985]: DEBUG oslo_vmware.api [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 925.503708] env[61985]: value = "task-936100" [ 925.503708] env[61985]: _type = "Task" [ 925.503708] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.513077] env[61985]: DEBUG oslo_vmware.api [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936100, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.524438] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936097, 'name': CreateVM_Task, 'duration_secs': 0.35345} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.524653] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 925.525631] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.525785] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.526218] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.526700] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a022af0b-093c-4d98-b228-2aede90501d4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.531411] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 925.531411] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522220d1-3ee1-6ca9-11eb-4ba97989e9f2" [ 925.531411] env[61985]: _type = "Task" [ 925.531411] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.542053] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522220d1-3ee1-6ca9-11eb-4ba97989e9f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.806327] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936098, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097246} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.806905] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.808092] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c20f46-8400-4c9c-9b2f-008009fe78b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.819811] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936099, 'name': Rename_Task, 'duration_secs': 0.168484} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.830990] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 925.840041] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 87006fbf-ea90-4d9a-88af-001de424ac14/87006fbf-ea90-4d9a-88af-001de424ac14.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.840332] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5036f68-f90f-4743-bdbf-f0230d8cc13c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.842550] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-375df6d8-d72f-4275-bc7a-80b7f33eb1ae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.864090] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Successfully updated port: 6ee77311-001c-4c2f-afad-e75094a76022 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.868424] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 925.868424] env[61985]: value = "task-936102" [ 925.868424] env[61985]: _type = "Task" [ 925.868424] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.868679] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 925.868679] env[61985]: value = "task-936101" [ 925.868679] env[61985]: _type = "Task" [ 925.868679] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.889065] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936102, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.889403] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936101, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.896963] env[61985]: DEBUG nova.compute.utils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 925.897204] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 925.897877] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 925.946712] env[61985]: DEBUG nova.policy [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3d8fab40f554eb3a4804a4291686146', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c06fc8ffa004e00be0a1991d2e082e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 925.953627] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6a2c3890-e7ae-4e0c-a311-26187b6c868c tempest-InstanceActionsTestJSON-1449424502 tempest-InstanceActionsTestJSON-1449424502-project-member] Lock "5b3a8ba3-48cb-472a-ab45-4e51af68812c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.165s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.954710] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d8f14a3-c5ab-43d4-9aa5-6334b931bdb1 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.287s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.958666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 23.759s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.958666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "c64056d8-0c22-48db-bf4a-a60f468bdad0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.958666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.958666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.958666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f59fc03-6264-4bf8-9128-c14e2e39d7a4 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "986c576e-be02-48ac-b24c-72edccab25be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.370s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.959810] env[61985]: INFO nova.compute.manager [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Terminating instance [ 925.961476] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.961743] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquired lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.961843] env[61985]: DEBUG nova.network.neutron [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 926.015904] env[61985]: DEBUG oslo_vmware.api [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936100, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.040402] env[61985]: DEBUG nova.objects.instance [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lazy-loading 'flavor' on Instance uuid aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.049708] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522220d1-3ee1-6ca9-11eb-4ba97989e9f2, 'name': SearchDatastore_Task, 'duration_secs': 0.015255} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.053034] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.053034] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.053236] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.053344] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.053561] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.054907] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fa10583-9c29-48a1-9327-0bb8547b30bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.066244] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.066244] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 926.066244] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1df01422-f32b-4e8d-a6bb-877d6d6c277a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.072190] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 926.072190] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5227f92b-7027-6ea0-445f-9e758c99f794" [ 926.072190] env[61985]: _type = "Task" [ 926.072190] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.081458] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5227f92b-7027-6ea0-445f-9e758c99f794, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.127780] env[61985]: DEBUG nova.network.neutron [req-0bb2fe08-abf3-46ed-92ef-14e2be85ad75 req-c372b531-a398-4487-838a-502a861842d3 service nova] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Updated VIF entry in instance network info cache for port 41b487bd-afda-4a0c-be1a-056e516b010c. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 926.128205] env[61985]: DEBUG nova.network.neutron [req-0bb2fe08-abf3-46ed-92ef-14e2be85ad75 req-c372b531-a398-4487-838a-502a861842d3 service nova] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Updating instance_info_cache with network_info: [{"id": "41b487bd-afda-4a0c-be1a-056e516b010c", "address": "fa:16:3e:a0:47:19", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41b487bd-af", "ovs_interfaceid": "41b487bd-afda-4a0c-be1a-056e516b010c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.370196] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "refresh_cache-62824eac-9412-466a-abcf-1010f6a829e4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.370553] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "refresh_cache-62824eac-9412-466a-abcf-1010f6a829e4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.370553] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 926.387352] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936102, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.393230] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936101, 'name': PowerOnVM_Task} progress is 70%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.401111] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 926.466039] env[61985]: DEBUG nova.compute.utils [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Can not refresh info_cache because instance was not found {{(pid=61985) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 926.487590] env[61985]: DEBUG nova.network.neutron [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 926.504932] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Successfully created port: 57bd4c36-e5e2-43f4-b8e5-7b574b97ba90 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.521602] env[61985]: DEBUG oslo_vmware.api [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936100, 'name': ReconfigVM_Task, 'duration_secs': 0.963201} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.521908] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Reconfigured VM instance instance-0000003a to attach disk [datastore1] volume-78b9d30c-3370-4505-b9bf-60bae00d9eda/volume-78b9d30c-3370-4505-b9bf-60bae00d9eda.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.528459] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b29dac63-c62e-4dbb-b17b-fa18784cb075 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.549088] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fae646a9-ab58-4f2b-b7e8-c79681716d9b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.917s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.559266] env[61985]: DEBUG oslo_vmware.api [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 926.559266] env[61985]: value = "task-936104" [ 926.559266] env[61985]: _type = "Task" [ 926.559266] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.569964] env[61985]: DEBUG oslo_vmware.api [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936104, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.589822] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5227f92b-7027-6ea0-445f-9e758c99f794, 'name': SearchDatastore_Task, 'duration_secs': 0.017021} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.590878] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42143775-82e5-4b93-b269-e69ece05c59a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.596982] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 926.596982] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52319f8a-fe9e-ad89-fff2-17fbe420885e" [ 926.596982] env[61985]: _type = "Task" [ 926.596982] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.609688] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52319f8a-fe9e-ad89-fff2-17fbe420885e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.618977] env[61985]: DEBUG nova.network.neutron [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.631946] env[61985]: DEBUG oslo_concurrency.lockutils [req-0bb2fe08-abf3-46ed-92ef-14e2be85ad75 req-c372b531-a398-4487-838a-502a861842d3 service nova] Releasing lock "refresh_cache-5c2a7072-5b14-4ec7-8060-4e21eea927e3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.829597] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64cc564f-d12e-4420-9113-8978d111639b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.841899] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7408a31-9595-43f8-b00e-d4c99deaec3a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.882738] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3bda40-ba48-4e89-bb81-1547cafc980f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.897204] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936102, 'name': ReconfigVM_Task, 'duration_secs': 0.7066} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.907413] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 87006fbf-ea90-4d9a-88af-001de424ac14/87006fbf-ea90-4d9a-88af-001de424ac14.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.908025] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936101, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.917404] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69da6405-01fd-431a-8189-3ef0893f5e63 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.917404] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b79c8d-d5d4-4b04-b08d-0317e6e1c25d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.937743] env[61985]: DEBUG nova.compute.provider_tree [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.939178] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 926.939178] env[61985]: value = "task-936105" [ 926.939178] env[61985]: _type = "Task" [ 926.939178] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.940386] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 926.955597] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936105, 'name': Rename_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.049675] env[61985]: DEBUG nova.compute.manager [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.050677] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de08ce3-4b6b-4f64-bc74-5a1ab079f35a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.071680] env[61985]: DEBUG oslo_vmware.api [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936104, 'name': ReconfigVM_Task, 'duration_secs': 0.196611} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.072038] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211408', 'volume_id': '78b9d30c-3370-4505-b9bf-60bae00d9eda', 'name': 'volume-78b9d30c-3370-4505-b9bf-60bae00d9eda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1f284789-1e7b-4e9f-9670-34e8e25cd797', 'attached_at': '', 'detached_at': '', 'volume_id': '78b9d30c-3370-4505-b9bf-60bae00d9eda', 'serial': '78b9d30c-3370-4505-b9bf-60bae00d9eda'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 927.112312] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52319f8a-fe9e-ad89-fff2-17fbe420885e, 'name': SearchDatastore_Task, 'duration_secs': 0.016873} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.112312] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.112493] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 5c2a7072-5b14-4ec7-8060-4e21eea927e3/5c2a7072-5b14-4ec7-8060-4e21eea927e3.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 927.112780] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a820184b-2e7b-4597-b1da-c67587884fbd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.120544] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 927.120544] env[61985]: value = "task-936106" [ 927.120544] env[61985]: _type = "Task" [ 927.120544] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.121036] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Releasing lock "refresh_cache-c64056d8-0c22-48db-bf4a-a60f468bdad0" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.121440] env[61985]: DEBUG nova.compute.manager [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 927.121639] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 927.125617] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2fc45ca-10be-4d3d-a053-60e121287553 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.133431] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936106, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.137716] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809a989b-7fb3-44c3-8471-ed11334c0fd0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.154817] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Updating instance_info_cache with network_info: [{"id": "6ee77311-001c-4c2f-afad-e75094a76022", "address": "fa:16:3e:89:0b:f4", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ee77311-00", "ovs_interfaceid": "6ee77311-001c-4c2f-afad-e75094a76022", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.180285] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c64056d8-0c22-48db-bf4a-a60f468bdad0 could not be found. [ 927.180556] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 927.180757] env[61985]: INFO nova.compute.manager [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Took 0.06 seconds to destroy the instance on the hypervisor. [ 927.181042] env[61985]: DEBUG oslo.service.loopingcall [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.182011] env[61985]: DEBUG nova.compute.manager [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.182160] env[61985]: DEBUG nova.network.neutron [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 927.200825] env[61985]: DEBUG nova.network.neutron [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 927.327162] env[61985]: DEBUG oslo_concurrency.lockutils [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.327478] env[61985]: DEBUG oslo_concurrency.lockutils [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.328066] env[61985]: DEBUG oslo_concurrency.lockutils [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.328066] env[61985]: DEBUG oslo_concurrency.lockutils [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.328339] env[61985]: DEBUG oslo_concurrency.lockutils [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.330806] env[61985]: INFO nova.compute.manager [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Terminating instance [ 927.332528] env[61985]: DEBUG nova.compute.manager [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 927.332734] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 927.335022] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79fd8b02-fda6-4a41-96d7-662c3bba6041 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.343115] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 927.343115] env[61985]: value = "task-936107" [ 927.343115] env[61985]: _type = "Task" [ 927.343115] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.348967] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936107, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.385892] env[61985]: DEBUG oslo_vmware.api [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936101, 'name': PowerOnVM_Task, 'duration_secs': 1.285556} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.386360] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 927.386561] env[61985]: INFO nova.compute.manager [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Took 10.54 seconds to spawn the instance on the hypervisor. [ 927.386826] env[61985]: DEBUG nova.compute.manager [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.387861] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2bdbe0-ebd6-48ef-a157-9e4a19bed469 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.422201] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 927.443613] env[61985]: DEBUG nova.scheduler.client.report [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.461416] env[61985]: DEBUG nova.compute.manager [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Received event network-vif-plugged-6ee77311-001c-4c2f-afad-e75094a76022 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 927.461750] env[61985]: DEBUG oslo_concurrency.lockutils [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] Acquiring lock "62824eac-9412-466a-abcf-1010f6a829e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.462047] env[61985]: DEBUG oslo_concurrency.lockutils [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] Lock "62824eac-9412-466a-abcf-1010f6a829e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.462707] env[61985]: DEBUG oslo_concurrency.lockutils [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] Lock "62824eac-9412-466a-abcf-1010f6a829e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.462707] env[61985]: DEBUG nova.compute.manager [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] No waiting events found dispatching network-vif-plugged-6ee77311-001c-4c2f-afad-e75094a76022 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 927.462707] env[61985]: WARNING nova.compute.manager [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Received unexpected event network-vif-plugged-6ee77311-001c-4c2f-afad-e75094a76022 for instance with vm_state building and task_state spawning. [ 927.462880] env[61985]: DEBUG nova.compute.manager [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Received event network-changed-6ee77311-001c-4c2f-afad-e75094a76022 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 927.462932] env[61985]: DEBUG nova.compute.manager [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Refreshing instance network info cache due to event network-changed-6ee77311-001c-4c2f-afad-e75094a76022. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 927.463151] env[61985]: DEBUG oslo_concurrency.lockutils [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] Acquiring lock "refresh_cache-62824eac-9412-466a-abcf-1010f6a829e4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.466669] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.466919] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.467130] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.467287] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.467459] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.467684] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.467905] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.468079] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.468256] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.468429] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.468634] env[61985]: DEBUG nova.virt.hardware [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.470879] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73bc2b52-5673-4c4a-96df-56ac43a65956 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.481583] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936105, 'name': Rename_Task, 'duration_secs': 0.211609} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.484583] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 927.484797] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f1d54d2-2a8d-4cc6-808f-1aa6d792403a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.487854] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb186a9-b67e-4ed5-aa1d-839f9e19581f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.498654] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 927.498654] env[61985]: value = "task-936108" [ 927.498654] env[61985]: _type = "Task" [ 927.498654] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.519356] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936108, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.568176] env[61985]: INFO nova.compute.manager [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] instance snapshotting [ 927.575733] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23ddf82-ca4f-4867-b128-0a76ebb943a9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.608442] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22442897-3a0c-474a-8fd1-9acc03b50a97 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.634806] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936106, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.657883] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "refresh_cache-62824eac-9412-466a-abcf-1010f6a829e4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.658634] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Instance network_info: |[{"id": "6ee77311-001c-4c2f-afad-e75094a76022", "address": "fa:16:3e:89:0b:f4", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ee77311-00", "ovs_interfaceid": "6ee77311-001c-4c2f-afad-e75094a76022", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 927.658969] env[61985]: DEBUG oslo_concurrency.lockutils [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] Acquired lock "refresh_cache-62824eac-9412-466a-abcf-1010f6a829e4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.659064] env[61985]: DEBUG nova.network.neutron [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Refreshing network info cache for port 6ee77311-001c-4c2f-afad-e75094a76022 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 927.661586] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:0b:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ee77311-001c-4c2f-afad-e75094a76022', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 927.673663] env[61985]: DEBUG oslo.service.loopingcall [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.679017] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 927.679112] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-438f81ed-d9d8-4e2c-8e7e-6be7fa159a1f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.704459] env[61985]: DEBUG nova.network.neutron [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.708928] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.708928] env[61985]: value = "task-936109" [ 927.708928] env[61985]: _type = "Task" [ 927.708928] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.721847] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936109, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.736105] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "3694e20c-ce37-4097-9991-8a06f38b2734" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.736503] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "3694e20c-ce37-4097-9991-8a06f38b2734" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.736768] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "3694e20c-ce37-4097-9991-8a06f38b2734-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.737020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "3694e20c-ce37-4097-9991-8a06f38b2734-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.737230] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "3694e20c-ce37-4097-9991-8a06f38b2734-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.743061] env[61985]: INFO nova.compute.manager [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Terminating instance [ 927.745537] env[61985]: DEBUG nova.compute.manager [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 927.745848] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 927.746935] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8f2694-4ee8-48e8-b001-c1ca8bcd2892 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.757928] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 927.758297] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dba2b6c5-7ae5-4d91-bb04-87732ba0be19 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.765567] env[61985]: DEBUG oslo_vmware.api [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 927.765567] env[61985]: value = "task-936110" [ 927.765567] env[61985]: _type = "Task" [ 927.765567] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.779232] env[61985]: DEBUG oslo_vmware.api [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-936110, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.853283] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936107, 'name': PowerOffVM_Task, 'duration_secs': 0.451537} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.853590] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 927.853804] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 927.854047] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211407', 'volume_id': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'name': 'volume-6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'serial': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 927.854874] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2198ba26-7878-448e-828f-c80772ee5014 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.877792] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed7903e-88b4-4bb4-aa01-eea2fbbaeb8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.888370] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4431d0-2887-46e7-b946-7900b06fa1d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.917751] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4297fe4e-9724-4d03-a977-55101075bd96 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.930150] env[61985]: INFO nova.compute.manager [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Took 47.05 seconds to build instance. [ 927.952993] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.953813] env[61985]: DEBUG nova.compute.manager [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 927.958155] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] The volume has not been displaced from its original location: [datastore1] volume-6c7f9ac2-d6bd-479b-86cb-5641a612d6bb/volume-6c7f9ac2-d6bd-479b-86cb-5641a612d6bb.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 927.967472] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Reconfiguring VM instance instance-00000045 to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 927.968499] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 26.707s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.970094] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b0a2647-f9b4-42f8-9120-49e729faee34 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.989292] env[61985]: DEBUG nova.network.neutron [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Updated VIF entry in instance network info cache for port 6ee77311-001c-4c2f-afad-e75094a76022. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 927.989761] env[61985]: DEBUG nova.network.neutron [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Updating instance_info_cache with network_info: [{"id": "6ee77311-001c-4c2f-afad-e75094a76022", "address": "fa:16:3e:89:0b:f4", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ee77311-00", "ovs_interfaceid": "6ee77311-001c-4c2f-afad-e75094a76022", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.993892] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 927.993892] env[61985]: value = "task-936111" [ 927.993892] env[61985]: _type = "Task" [ 927.993892] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.004010] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936111, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.022879] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936108, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.126300] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 928.130248] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7d38cf3c-62fd-45ad-aa16-7875cc8aee4a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.140390] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936106, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550489} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.141447] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 5c2a7072-5b14-4ec7-8060-4e21eea927e3/5c2a7072-5b14-4ec7-8060-4e21eea927e3.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 928.141704] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.142021] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 928.142021] env[61985]: value = "task-936112" [ 928.142021] env[61985]: _type = "Task" [ 928.142021] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.142271] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51dc415d-ec8b-49f7-b816-1348fccad4ac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.153723] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936112, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.155629] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 928.155629] env[61985]: value = "task-936113" [ 928.155629] env[61985]: _type = "Task" [ 928.155629] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.163386] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936113, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.207071] env[61985]: INFO nova.compute.manager [-] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Took 1.02 seconds to deallocate network for instance. [ 928.222747] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936109, 'name': CreateVM_Task, 'duration_secs': 0.494345} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.222928] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 928.223719] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.223997] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.224246] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.224522] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9dcdb7a-c6ca-4dcb-b9f9-728bd6bd4eb4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.229643] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 928.229643] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5251b3ec-94a9-3362-8fd0-d48856f1b64f" [ 928.229643] env[61985]: _type = "Task" [ 928.229643] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.237849] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5251b3ec-94a9-3362-8fd0-d48856f1b64f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.275429] env[61985]: DEBUG nova.objects.instance [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lazy-loading 'flavor' on Instance uuid 1f284789-1e7b-4e9f-9670-34e8e25cd797 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.277489] env[61985]: DEBUG oslo_vmware.api [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-936110, 'name': PowerOffVM_Task, 'duration_secs': 0.250925} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.278092] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 928.278899] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 928.278899] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35ab6174-d87c-42b0-b170-dc0c4ffc1a2a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.340360] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 928.340635] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 928.340864] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Deleting the datastore file [datastore1] 3694e20c-ce37-4097-9991-8a06f38b2734 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 928.341204] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78f0c7f7-bbdc-4739-ae78-5af71cbd54ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.348759] env[61985]: DEBUG oslo_vmware.api [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for the task: (returnval){ [ 928.348759] env[61985]: value = "task-936115" [ 928.348759] env[61985]: _type = "Task" [ 928.348759] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.359258] env[61985]: DEBUG oslo_vmware.api [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-936115, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.434310] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5594f781-39f6-424a-b6e2-256c8293b148 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.811s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.469150] env[61985]: DEBUG nova.compute.utils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 928.474019] env[61985]: DEBUG nova.compute.manager [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 928.474019] env[61985]: DEBUG nova.network.neutron [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 928.488627] env[61985]: INFO nova.compute.claims [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.495123] env[61985]: DEBUG oslo_concurrency.lockutils [req-fc18f1b6-b4c6-45cd-a08e-90bb8cd29ea6 req-25754129-75b3-48b9-9428-f7264e6f42b6 service nova] Releasing lock "refresh_cache-62824eac-9412-466a-abcf-1010f6a829e4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.504380] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936111, 'name': ReconfigVM_Task, 'duration_secs': 0.238661} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.504660] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Reconfigured VM instance instance-00000045 to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 928.509765] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ddd678d-661e-4897-a442-ccf057a934ef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.524418] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 928.524647] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211414', 'volume_id': '774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'name': 'volume-774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce757dda-f58a-47b3-b319-e6b01f05b20b', 'attached_at': '', 'detached_at': '', 'volume_id': '774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'serial': '774d6c9b-3a7e-48dc-836f-82cec83d6e53'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 928.528664] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415ef286-1cda-4d73-b645-51d2a7d7103a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.537818] env[61985]: DEBUG oslo_vmware.api [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936108, 'name': PowerOnVM_Task, 'duration_secs': 0.716069} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.549054] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 928.549396] env[61985]: INFO nova.compute.manager [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Took 9.19 seconds to spawn the instance on the hypervisor. [ 928.549610] env[61985]: DEBUG nova.compute.manager [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.549960] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 928.549960] env[61985]: value = "task-936116" [ 928.549960] env[61985]: _type = "Task" [ 928.549960] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.551148] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a237ce-1917-4c94-98e4-1f51ae7b5465 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.554410] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10660615-2369-4bba-bdc9-6b811b1d7801 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.572321] env[61985]: DEBUG nova.policy [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '149807bc94cc4cddb59e4dad0bf1099c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8015ef21eaf4a0bb19063e261064cdc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 928.587125] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936116, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.596778] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] volume-774d6c9b-3a7e-48dc-836f-82cec83d6e53/volume-774d6c9b-3a7e-48dc-836f-82cec83d6e53.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.596778] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ac11190-f589-481f-805b-f8881b8aba55 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.615076] env[61985]: DEBUG oslo_vmware.api [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 928.615076] env[61985]: value = "task-936117" [ 928.615076] env[61985]: _type = "Task" [ 928.615076] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.625856] env[61985]: DEBUG oslo_vmware.api [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936117, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.653855] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936112, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.668062] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936113, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0803} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.668062] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.668062] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254dc0fe-4729-40db-a9e8-a36d39b03b1b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.691248] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 5c2a7072-5b14-4ec7-8060-4e21eea927e3/5c2a7072-5b14-4ec7-8060-4e21eea927e3.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.692121] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21f27e69-513c-47aa-912d-d34d42e5f3c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.713467] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 928.713467] env[61985]: value = "task-936118" [ 928.713467] env[61985]: _type = "Task" [ 928.713467] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.714053] env[61985]: INFO nova.compute.manager [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance disappeared during terminate [ 928.714578] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ba811ee-bcc5-4e96-afc0-60bc898358d4 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "c64056d8-0c22-48db-bf4a-a60f468bdad0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.758s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.725021] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936118, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.743880] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5251b3ec-94a9-3362-8fd0-d48856f1b64f, 'name': SearchDatastore_Task, 'duration_secs': 0.012567} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.744940] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.745259] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.745521] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.745689] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.745895] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.746509] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3b7e8a4-291c-40f7-b6a7-b93d1b281aaf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.757189] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.757556] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 928.758685] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8928ec84-2b92-4b6c-8f1e-b94dc1232e85 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.766642] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 928.766642] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529310f4-def3-6f3f-e4dd-1bbf3296c132" [ 928.766642] env[61985]: _type = "Task" [ 928.766642] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.781642] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529310f4-def3-6f3f-e4dd-1bbf3296c132, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.782202] env[61985]: DEBUG oslo_concurrency.lockutils [None req-20b6064c-1369-47bb-bd93-ca5d94676ac7 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.025s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.860929] env[61985]: DEBUG oslo_vmware.api [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Task: {'id': task-936115, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.354399} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.861224] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.861414] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 928.861741] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 928.861824] env[61985]: INFO nova.compute.manager [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Took 1.12 seconds to destroy the instance on the hypervisor. [ 928.862049] env[61985]: DEBUG oslo.service.loopingcall [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.862241] env[61985]: DEBUG nova.compute.manager [-] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 928.862339] env[61985]: DEBUG nova.network.neutron [-] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 928.939975] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Successfully updated port: 57bd4c36-e5e2-43f4-b8e5-7b574b97ba90 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.972374] env[61985]: DEBUG nova.compute.manager [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 928.995210] env[61985]: INFO nova.compute.resource_tracker [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating resource usage from migration 9a221221-2558-443c-8a47-3d0ca57efb01 [ 929.070870] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936116, 'name': ReconfigVM_Task, 'duration_secs': 0.226918} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.073502] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211407', 'volume_id': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'name': 'volume-6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb', 'serial': '6c7f9ac2-d6bd-479b-86cb-5641a612d6bb'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 929.073849] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 929.074935] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b64fe7-ceac-408c-9d7e-3c4bbb563da7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.082101] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 929.082338] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7a27caf-4735-4845-9191-2d46c3d9c8f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.107233] env[61985]: INFO nova.compute.manager [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Took 44.88 seconds to build instance. [ 929.130636] env[61985]: DEBUG oslo_vmware.api [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.151374] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 929.151663] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 929.151894] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleting the datastore file [datastore1] aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.153465] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2aebaef5-d435-4e39-ab73-75df0d317d4b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.161159] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936112, 'name': CreateSnapshot_Task, 'duration_secs': 0.922034} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.162203] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 929.163180] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db15d6e-d969-4118-a73f-e4bcdd13fbf0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.168902] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 929.168902] env[61985]: value = "task-936120" [ 929.168902] env[61985]: _type = "Task" [ 929.168902] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.186774] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936120, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.201650] env[61985]: DEBUG nova.network.neutron [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Successfully created port: 1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.224374] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936118, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.283039] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529310f4-def3-6f3f-e4dd-1bbf3296c132, 'name': SearchDatastore_Task, 'duration_secs': 0.017804} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.283039] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-717e715d-a2b0-42e6-ab3d-e5bcf71ba424 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.290084] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 929.290084] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f11354-d947-2cf8-7f46-0b17cc3aafbd" [ 929.290084] env[61985]: _type = "Task" [ 929.290084] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.309309] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f11354-d947-2cf8-7f46-0b17cc3aafbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.445669] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "refresh_cache-cd8e2eaf-3c13-452b-b2e6-8107f0219378" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.445904] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "refresh_cache-cd8e2eaf-3c13-452b-b2e6-8107f0219378" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.445948] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 929.460326] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b54d7f-6fdd-40bd-ad31-fe5a1ebaa0ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.469093] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14be1012-c407-4cdc-9894-60c47b198c7f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.506697] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4a1c38-fec0-40e6-b5e3-5e61e1f5be57 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.512843] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d73b79-45b4-487d-b546-7ed2356f8dd9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.529077] env[61985]: DEBUG nova.compute.provider_tree [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.583556] env[61985]: DEBUG nova.compute.manager [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Received event network-vif-plugged-57bd4c36-e5e2-43f4-b8e5-7b574b97ba90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 929.583796] env[61985]: DEBUG oslo_concurrency.lockutils [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] Acquiring lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.584057] env[61985]: DEBUG oslo_concurrency.lockutils [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] Lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.584250] env[61985]: DEBUG oslo_concurrency.lockutils [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] Lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.584431] env[61985]: DEBUG nova.compute.manager [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] No waiting events found dispatching network-vif-plugged-57bd4c36-e5e2-43f4-b8e5-7b574b97ba90 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.584604] env[61985]: WARNING nova.compute.manager [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Received unexpected event network-vif-plugged-57bd4c36-e5e2-43f4-b8e5-7b574b97ba90 for instance with vm_state building and task_state spawning. [ 929.584774] env[61985]: DEBUG nova.compute.manager [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Received event network-changed-57bd4c36-e5e2-43f4-b8e5-7b574b97ba90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 929.584935] env[61985]: DEBUG nova.compute.manager [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Refreshing instance network info cache due to event network-changed-57bd4c36-e5e2-43f4-b8e5-7b574b97ba90. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 929.585130] env[61985]: DEBUG oslo_concurrency.lockutils [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] Acquiring lock "refresh_cache-cd8e2eaf-3c13-452b-b2e6-8107f0219378" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.610762] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1784935f-31c8-41a3-aac3-a7140737d6ae tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "87006fbf-ea90-4d9a-88af-001de424ac14" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.014s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.627899] env[61985]: DEBUG oslo_vmware.api [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936117, 'name': ReconfigVM_Task, 'duration_secs': 0.611785} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.628225] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Reconfigured VM instance instance-0000003d to attach disk [datastore1] volume-774d6c9b-3a7e-48dc-836f-82cec83d6e53/volume-774d6c9b-3a7e-48dc-836f-82cec83d6e53.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.633241] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e463a66a-08d8-49c0-ac7e-c47a802bb200 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.666044] env[61985]: DEBUG oslo_vmware.api [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 929.666044] env[61985]: value = "task-936121" [ 929.666044] env[61985]: _type = "Task" [ 929.666044] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.681308] env[61985]: DEBUG oslo_vmware.api [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936121, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.692271] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 929.692906] env[61985]: DEBUG oslo_vmware.api [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936120, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.359493} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.693279] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fa64daa1-d319-4cbc-abd4-f531eee11347 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.697228] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.697440] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 929.697669] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 929.697874] env[61985]: INFO nova.compute.manager [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Took 2.37 seconds to destroy the instance on the hypervisor. [ 929.698142] env[61985]: DEBUG oslo.service.loopingcall [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.698360] env[61985]: DEBUG nova.compute.manager [-] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 929.698446] env[61985]: DEBUG nova.network.neutron [-] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 929.707110] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 929.707110] env[61985]: value = "task-936122" [ 929.707110] env[61985]: _type = "Task" [ 929.707110] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.718143] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936122, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.730527] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936118, 'name': ReconfigVM_Task, 'duration_secs': 0.564754} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.730817] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 5c2a7072-5b14-4ec7-8060-4e21eea927e3/5c2a7072-5b14-4ec7-8060-4e21eea927e3.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.731664] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3011990-1048-40c6-b870-cf6fc3dd2265 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.743034] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 929.743034] env[61985]: value = "task-936123" [ 929.743034] env[61985]: _type = "Task" [ 929.743034] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.755645] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936123, 'name': Rename_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.806491] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f11354-d947-2cf8-7f46-0b17cc3aafbd, 'name': SearchDatastore_Task, 'duration_secs': 0.031876} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.807824] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.808473] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 62824eac-9412-466a-abcf-1010f6a829e4/62824eac-9412-466a-abcf-1010f6a829e4.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 929.809124] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2617406a-4a05-4a90-bec1-d1ed2b7fd719 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.816282] env[61985]: DEBUG nova.compute.manager [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Stashing vm_state: active {{(pid=61985) _prep_resize /opt/stack/nova/nova/compute/manager.py:5630}} [ 929.822556] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 929.822556] env[61985]: value = "task-936124" [ 929.822556] env[61985]: _type = "Task" [ 929.822556] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.836373] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.879599] env[61985]: DEBUG nova.network.neutron [-] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.006166] env[61985]: DEBUG nova.compute.manager [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 930.019328] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 930.038808] env[61985]: DEBUG nova.scheduler.client.report [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.056939] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.056939] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.056939] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.056939] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.056939] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.056939] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.056939] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.056939] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.056939] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.056939] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.057370] env[61985]: DEBUG nova.virt.hardware [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.057995] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24af4d37-5eff-4cfc-9734-69b0012ec361 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.067881] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c314fec9-524e-48b4-bf91-df22d3e8a823 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.118742] env[61985]: DEBUG oslo_vmware.rw_handles [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d818f8-8f6f-9333-ace3-de745a5aaf1e/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 930.122736] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae512d9-767b-48e3-9c3b-3d00f3c19102 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.130311] env[61985]: DEBUG oslo_vmware.rw_handles [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d818f8-8f6f-9333-ace3-de745a5aaf1e/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 930.130432] env[61985]: ERROR oslo_vmware.rw_handles [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d818f8-8f6f-9333-ace3-de745a5aaf1e/disk-0.vmdk due to incomplete transfer. [ 930.130791] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b2199370-59ec-4136-8b40-175ed9a75c44 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.138669] env[61985]: DEBUG oslo_vmware.rw_handles [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d818f8-8f6f-9333-ace3-de745a5aaf1e/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 930.138894] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Uploaded image 71734001-1b38-49c6-84ab-bdbcf8ca062f to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 930.142281] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 930.142666] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-11f1423f-c688-4c43-806a-53a04be06027 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.152539] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 930.152539] env[61985]: value = "task-936125" [ 930.152539] env[61985]: _type = "Task" [ 930.152539] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.169374] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936125, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.182308] env[61985]: DEBUG oslo_vmware.api [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936121, 'name': ReconfigVM_Task, 'duration_secs': 0.166926} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.182471] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211414', 'volume_id': '774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'name': 'volume-774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce757dda-f58a-47b3-b319-e6b01f05b20b', 'attached_at': '', 'detached_at': '', 'volume_id': '774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'serial': '774d6c9b-3a7e-48dc-836f-82cec83d6e53'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 930.221231] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936122, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.256276] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936123, 'name': Rename_Task, 'duration_secs': 0.196033} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.256755] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 930.257164] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ebe876fb-dee3-4f6f-96b4-82e1c3562c5e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.265012] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 930.265012] env[61985]: value = "task-936126" [ 930.265012] env[61985]: _type = "Task" [ 930.265012] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.276626] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.277065] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.286028] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936126, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.341461] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936124, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.343177] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.383317] env[61985]: INFO nova.compute.manager [-] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Took 1.52 seconds to deallocate network for instance. [ 930.430257] env[61985]: DEBUG nova.network.neutron [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Updating instance_info_cache with network_info: [{"id": "57bd4c36-e5e2-43f4-b8e5-7b574b97ba90", "address": "fa:16:3e:71:f3:9a", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57bd4c36-e5", "ovs_interfaceid": "57bd4c36-e5e2-43f4-b8e5-7b574b97ba90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.550598] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.580s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.550598] env[61985]: INFO nova.compute.manager [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Migrating [ 930.557471] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.097s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.557471] env[61985]: DEBUG nova.objects.instance [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lazy-loading 'resources' on Instance uuid 54ba5cfd-185a-4c58-aa5e-83cc096a482e {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.665381] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936125, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.679266] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.679505] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.724149] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936122, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.777313] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936126, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.789829] env[61985]: INFO nova.compute.manager [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Detaching volume 2e7a203f-be54-4d0d-b6df-a5b2894cd4d4 [ 930.833838] env[61985]: INFO nova.virt.block_device [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Attempting to driver detach volume 2e7a203f-be54-4d0d-b6df-a5b2894cd4d4 from mountpoint /dev/sdb [ 930.834448] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 930.834448] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211401', 'volume_id': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'name': 'volume-2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1f284789-1e7b-4e9f-9670-34e8e25cd797', 'attached_at': '', 'detached_at': '', 'volume_id': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'serial': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 930.837188] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4551da-162d-4ad7-bca9-cc313e52c9c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.842416] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936124, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546243} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.842416] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 62824eac-9412-466a-abcf-1010f6a829e4/62824eac-9412-466a-abcf-1010f6a829e4.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 930.842886] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.843196] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2628d1bc-f622-4109-a465-bfbbad4c7e1d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.869055] env[61985]: DEBUG nova.network.neutron [-] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.878674] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fde0af9-d7a9-40da-8bd4-3466bf84b881 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.882259] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 930.882259] env[61985]: value = "task-936127" [ 930.882259] env[61985]: _type = "Task" [ 930.882259] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.890907] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.892652] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f331b42f-08ad-45b2-b796-104b0f44a302 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.898975] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936127, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.922596] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e213f6f-1f10-47b5-b3cd-6ffb8b920e36 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.941643] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "refresh_cache-cd8e2eaf-3c13-452b-b2e6-8107f0219378" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.941904] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Instance network_info: |[{"id": "57bd4c36-e5e2-43f4-b8e5-7b574b97ba90", "address": "fa:16:3e:71:f3:9a", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57bd4c36-e5", "ovs_interfaceid": "57bd4c36-e5e2-43f4-b8e5-7b574b97ba90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.943340] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] The volume has not been displaced from its original location: [datastore1] volume-2e7a203f-be54-4d0d-b6df-a5b2894cd4d4/volume-2e7a203f-be54-4d0d-b6df-a5b2894cd4d4.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 930.948522] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Reconfiguring VM instance instance-0000003a to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 930.949455] env[61985]: DEBUG oslo_concurrency.lockutils [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] Acquired lock "refresh_cache-cd8e2eaf-3c13-452b-b2e6-8107f0219378" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.949455] env[61985]: DEBUG nova.network.neutron [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Refreshing network info cache for port 57bd4c36-e5e2-43f4-b8e5-7b574b97ba90 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 930.950809] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:f3:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57bd4c36-e5e2-43f4-b8e5-7b574b97ba90', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.959349] env[61985]: DEBUG oslo.service.loopingcall [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.960116] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-356a1464-87b6-4cf4-9139-8661c89fb9d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.974958] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 930.979477] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7791b6d2-c2a9-464f-a60e-37e2fb546851 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.998830] env[61985]: DEBUG oslo_vmware.api [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 930.998830] env[61985]: value = "task-936128" [ 930.998830] env[61985]: _type = "Task" [ 930.998830] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.000288] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.000288] env[61985]: value = "task-936129" [ 931.000288] env[61985]: _type = "Task" [ 931.000288] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.011946] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936129, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.016342] env[61985]: DEBUG oslo_vmware.api [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936128, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.040369] env[61985]: DEBUG nova.compute.manager [req-0844bbfd-ba7b-48d1-839f-ada304a54933 req-272e16a5-c724-45ff-b16f-76333f202286 service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Received event network-vif-plugged-1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 931.040610] env[61985]: DEBUG oslo_concurrency.lockutils [req-0844bbfd-ba7b-48d1-839f-ada304a54933 req-272e16a5-c724-45ff-b16f-76333f202286 service nova] Acquiring lock "488d0b95-97a9-4193-af62-cc92caf99625-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.040830] env[61985]: DEBUG oslo_concurrency.lockutils [req-0844bbfd-ba7b-48d1-839f-ada304a54933 req-272e16a5-c724-45ff-b16f-76333f202286 service nova] Lock "488d0b95-97a9-4193-af62-cc92caf99625-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.041071] env[61985]: DEBUG oslo_concurrency.lockutils [req-0844bbfd-ba7b-48d1-839f-ada304a54933 req-272e16a5-c724-45ff-b16f-76333f202286 service nova] Lock "488d0b95-97a9-4193-af62-cc92caf99625-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.041355] env[61985]: DEBUG nova.compute.manager [req-0844bbfd-ba7b-48d1-839f-ada304a54933 req-272e16a5-c724-45ff-b16f-76333f202286 service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] No waiting events found dispatching network-vif-plugged-1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 931.041463] env[61985]: WARNING nova.compute.manager [req-0844bbfd-ba7b-48d1-839f-ada304a54933 req-272e16a5-c724-45ff-b16f-76333f202286 service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Received unexpected event network-vif-plugged-1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4 for instance with vm_state building and task_state spawning. [ 931.067073] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.067705] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.067705] env[61985]: DEBUG nova.network.neutron [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 931.166121] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936125, 'name': Destroy_Task, 'duration_secs': 0.95868} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.166433] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Destroyed the VM [ 931.166744] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 931.167127] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e43474c3-c1d9-4eac-bde0-b17407e1fe15 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.174090] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 931.174090] env[61985]: value = "task-936130" [ 931.174090] env[61985]: _type = "Task" [ 931.174090] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.182579] env[61985]: DEBUG nova.compute.manager [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 931.188672] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936130, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.219917] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936122, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.239042] env[61985]: DEBUG nova.network.neutron [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Successfully updated port: 1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 931.240897] env[61985]: DEBUG nova.objects.instance [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lazy-loading 'flavor' on Instance uuid ce757dda-f58a-47b3-b319-e6b01f05b20b {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.284373] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936126, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.384748] env[61985]: INFO nova.compute.manager [-] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Took 1.68 seconds to deallocate network for instance. [ 931.410245] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936127, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069763} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.411215] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.412148] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2c4ae9-ea67-4b45-b957-0f0aed58f8b7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.435159] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 62824eac-9412-466a-abcf-1010f6a829e4/62824eac-9412-466a-abcf-1010f6a829e4.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.438100] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a3c40bd-af99-4085-bcd7-2451aad44789 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.459713] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 931.459713] env[61985]: value = "task-936131" [ 931.459713] env[61985]: _type = "Task" [ 931.459713] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.473799] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936131, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.522638] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936129, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.523491] env[61985]: DEBUG oslo_vmware.api [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936128, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.654966] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c80cc2c-8f27-4e3c-aee6-7595d1a9ae57 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.673485] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92586ce6-8093-4de9-a3d4-93a66f9528b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.747906] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "refresh_cache-488d0b95-97a9-4193-af62-cc92caf99625" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.748155] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired lock "refresh_cache-488d0b95-97a9-4193-af62-cc92caf99625" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.748319] env[61985]: DEBUG nova.network.neutron [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 931.753060] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc21f815-a106-4466-8caa-b6c6ff3ad4f0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.764802] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936130, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.765435] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2b11169-3024-479a-983e-148c5ab6e1ac tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.881s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.777274] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936122, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.778931] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372c7417-6bab-40d9-9f7e-d9dd2ca5fa6c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.783716] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.787656] env[61985]: DEBUG oslo_vmware.api [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936126, 'name': PowerOnVM_Task, 'duration_secs': 1.017171} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.788326] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 931.788749] env[61985]: INFO nova.compute.manager [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Took 9.52 seconds to spawn the instance on the hypervisor. [ 931.788975] env[61985]: DEBUG nova.compute.manager [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.789792] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09474899-4566-498b-95b3-bf07a238caf9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.802217] env[61985]: DEBUG nova.compute.provider_tree [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 931.819996] env[61985]: DEBUG nova.compute.manager [req-dc288dfd-e75c-4d4c-90a9-ccb120b723b3 req-087fa444-27b7-4657-ab32-103c5124f43e service nova] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Received event network-vif-deleted-f9834461-44c6-4878-8da9-e4a918ad3f51 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 931.859182] env[61985]: DEBUG nova.network.neutron [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 931.861422] env[61985]: DEBUG nova.network.neutron [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Updated VIF entry in instance network info cache for port 57bd4c36-e5e2-43f4-b8e5-7b574b97ba90. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 931.862854] env[61985]: DEBUG nova.network.neutron [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Updating instance_info_cache with network_info: [{"id": "57bd4c36-e5e2-43f4-b8e5-7b574b97ba90", "address": "fa:16:3e:71:f3:9a", "network": {"id": "4bf127aa-5eae-4584-b856-78a2a1a6b7b8", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1491104370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c06fc8ffa004e00be0a1991d2e082e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57bd4c36-e5", "ovs_interfaceid": "57bd4c36-e5e2-43f4-b8e5-7b574b97ba90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.958280] env[61985]: INFO nova.compute.manager [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Took 0.58 seconds to detach 1 volumes for instance. [ 931.974590] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936131, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.039243] env[61985]: DEBUG oslo_vmware.api [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936128, 'name': ReconfigVM_Task, 'duration_secs': 0.61991} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.042526] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Reconfigured VM instance instance-0000003a to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 932.048141] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936129, 'name': CreateVM_Task, 'duration_secs': 0.824955} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.048141] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa7e82d4-f5af-4c64-8a1e-8271270e978b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.062199] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 932.067288] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.067460] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.067928] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 932.068873] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20458598-fdd9-4257-8e6b-1e2097d73b89 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.074651] env[61985]: DEBUG oslo_vmware.api [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 932.074651] env[61985]: value = "task-936132" [ 932.074651] env[61985]: _type = "Task" [ 932.074651] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.076613] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 932.076613] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521b5a02-f6af-3e8f-1f06-2fb6012f58d1" [ 932.076613] env[61985]: _type = "Task" [ 932.076613] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.095510] env[61985]: DEBUG oslo_vmware.api [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936132, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.099682] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521b5a02-f6af-3e8f-1f06-2fb6012f58d1, 'name': SearchDatastore_Task, 'duration_secs': 0.018141} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.099949] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.100210] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.100566] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.100737] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.100932] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.101263] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a3e17be-3ef4-49cd-a97b-2a30cc7572e6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.112232] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.112813] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 932.113223] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fa5cabb-56e1-4878-a6ac-a98ffe5ea628 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.119215] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 932.119215] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5267c515-6a5d-b74e-bf80-bffa037b2bb6" [ 932.119215] env[61985]: _type = "Task" [ 932.119215] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.127730] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5267c515-6a5d-b74e-bf80-bffa037b2bb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.193752] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936130, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.252852] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936122, 'name': CloneVM_Task, 'duration_secs': 2.132822} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.253388] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Created linked-clone VM from snapshot [ 932.254415] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73036769-6e0c-440d-b679-168e1610c2e4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.262791] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Uploading image b4b4e229-a7cf-45de-8912-47f4b206140b {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 932.286972] env[61985]: DEBUG oslo_vmware.rw_handles [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 932.286972] env[61985]: value = "vm-211418" [ 932.286972] env[61985]: _type = "VirtualMachine" [ 932.286972] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 932.287308] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8e95483a-d22e-4a4c-8eaa-250696b3d4d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.296253] env[61985]: DEBUG oslo_vmware.rw_handles [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lease: (returnval){ [ 932.296253] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5228c62c-9844-ee7f-9a81-04012ebb14d3" [ 932.296253] env[61985]: _type = "HttpNfcLease" [ 932.296253] env[61985]: } obtained for exporting VM: (result){ [ 932.296253] env[61985]: value = "vm-211418" [ 932.296253] env[61985]: _type = "VirtualMachine" [ 932.296253] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 932.296646] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the lease: (returnval){ [ 932.296646] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5228c62c-9844-ee7f-9a81-04012ebb14d3" [ 932.296646] env[61985]: _type = "HttpNfcLease" [ 932.296646] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 932.303234] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 932.303234] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5228c62c-9844-ee7f-9a81-04012ebb14d3" [ 932.303234] env[61985]: _type = "HttpNfcLease" [ 932.303234] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 932.328024] env[61985]: INFO nova.compute.manager [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Took 46.24 seconds to build instance. [ 932.342335] env[61985]: ERROR nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [req-945e7a2b-30aa-4a62-b313-4f667019aa69] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-945e7a2b-30aa-4a62-b313-4f667019aa69"}]} [ 932.358977] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 932.367178] env[61985]: DEBUG oslo_concurrency.lockutils [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] Releasing lock "refresh_cache-cd8e2eaf-3c13-452b-b2e6-8107f0219378" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.367473] env[61985]: DEBUG nova.compute.manager [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Received event network-vif-deleted-60c2bc37-3190-4dfb-8b71-fd6eb3c3949b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 932.368164] env[61985]: INFO nova.compute.manager [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Neutron deleted interface 60c2bc37-3190-4dfb-8b71-fd6eb3c3949b; detaching it from the instance and deleting it from the info cache [ 932.368164] env[61985]: DEBUG nova.network.neutron [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.373287] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 932.373505] env[61985]: DEBUG nova.compute.provider_tree [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.386281] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 932.393875] env[61985]: DEBUG nova.network.neutron [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Updating instance_info_cache with network_info: [{"id": "1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4", "address": "fa:16:3e:b3:5d:0f", "network": {"id": "ddf63d11-2633-4be8-b73f-c5ca846b420c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-709607861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8015ef21eaf4a0bb19063e261064cdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bfe0ade-ba", "ovs_interfaceid": "1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.405736] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 932.468683] env[61985]: DEBUG nova.network.neutron [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance_info_cache with network_info: [{"id": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "address": "fa:16:3e:12:45:ef", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aa18b7b-89", "ovs_interfaceid": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.469695] env[61985]: DEBUG oslo_concurrency.lockutils [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.475073] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936131, 'name': ReconfigVM_Task, 'duration_secs': 0.648384} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.477580] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 62824eac-9412-466a-abcf-1010f6a829e4/62824eac-9412-466a-abcf-1010f6a829e4.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.478923] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5656b53f-8b72-4d7d-b5e8-1622ba4b0930 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.486174] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 932.486174] env[61985]: value = "task-936134" [ 932.486174] env[61985]: _type = "Task" [ 932.486174] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.496911] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936134, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.585282] env[61985]: DEBUG oslo_vmware.api [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936132, 'name': ReconfigVM_Task, 'duration_secs': 0.19749} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.585612] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211401', 'volume_id': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'name': 'volume-2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1f284789-1e7b-4e9f-9670-34e8e25cd797', 'attached_at': '', 'detached_at': '', 'volume_id': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4', 'serial': '2e7a203f-be54-4d0d-b6df-a5b2894cd4d4'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 932.632159] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5267c515-6a5d-b74e-bf80-bffa037b2bb6, 'name': SearchDatastore_Task, 'duration_secs': 0.012504} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.632483] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87887849-fff1-4a6d-8692-1b23b4e95b93 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.638084] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 932.638084] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52450be7-3db8-e693-b50d-a893efab8750" [ 932.638084] env[61985]: _type = "Task" [ 932.638084] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.648680] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52450be7-3db8-e693-b50d-a893efab8750, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.696924] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936130, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.805893] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 932.805893] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5228c62c-9844-ee7f-9a81-04012ebb14d3" [ 932.805893] env[61985]: _type = "HttpNfcLease" [ 932.805893] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 932.815952] env[61985]: DEBUG oslo_vmware.rw_handles [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 932.815952] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5228c62c-9844-ee7f-9a81-04012ebb14d3" [ 932.815952] env[61985]: _type = "HttpNfcLease" [ 932.815952] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 932.816549] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbaad752-320a-4d01-b9c2-57b9542290c9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.828265] env[61985]: DEBUG oslo_vmware.rw_handles [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d3c40a-c43f-6b3d-a50b-2b96eb6c2649/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 932.829390] env[61985]: DEBUG oslo_vmware.rw_handles [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d3c40a-c43f-6b3d-a50b-2b96eb6c2649/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 932.833285] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e18e67cf-b4e3-495d-8306-660683fc26c7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.887s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.834337] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7158b930-1d8c-4f2f-9f62-180227f96d6e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.901632] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Releasing lock "refresh_cache-488d0b95-97a9-4193-af62-cc92caf99625" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.901632] env[61985]: DEBUG nova.compute.manager [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Instance network_info: |[{"id": "1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4", "address": "fa:16:3e:b3:5d:0f", "network": {"id": "ddf63d11-2633-4be8-b73f-c5ca846b420c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-709607861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8015ef21eaf4a0bb19063e261064cdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bfe0ade-ba", "ovs_interfaceid": "1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 932.904206] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bda5b586-63f4-40ed-b21a-2f13a2129ed8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.906360] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:5d:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1470a3f8-be8a-4339-8a6f-9519366f32e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 932.914267] env[61985]: DEBUG oslo.service.loopingcall [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.915761] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 932.916746] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22a8ac7-c870-4848-bc0b-5bb87672aa09 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.921952] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35a87596-5a7c-4210-a073-99f7f77d3dd0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.945499] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aba151c-a871-4d7e-b3ae-a48d11815f9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.985331] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 932.985331] env[61985]: value = "task-936135" [ 932.985331] env[61985]: _type = "Task" [ 932.985331] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.985945] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.991643] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ad53f0-59de-4b65-8341-1f0de75f7df7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.002150] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1baece2a-0748-4fd1-a056-eca4590bd7b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.019059] env[61985]: DEBUG nova.compute.manager [req-0ceda803-53b3-45df-b651-239580af20fd req-d72e7765-66f0-40b2-acf7-319abef7be7d service nova] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Detach interface failed, port_id=60c2bc37-3190-4dfb-8b71-fd6eb3c3949b, reason: Instance 3694e20c-ce37-4097-9991-8a06f38b2734 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 933.036244] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad76435-5476-4604-a1f1-7f11be1a7138 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.039569] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936135, 'name': CreateVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.039890] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936134, 'name': Rename_Task, 'duration_secs': 0.248347} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.043764] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 933.044259] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3b35cdb-c69a-499a-9306-3e3ecce2f3ea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.056799] env[61985]: DEBUG nova.compute.provider_tree [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 933.067503] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 933.067503] env[61985]: value = "task-936136" [ 933.067503] env[61985]: _type = "Task" [ 933.067503] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.073325] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936136, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.153887] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52450be7-3db8-e693-b50d-a893efab8750, 'name': SearchDatastore_Task, 'duration_secs': 0.028456} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.154831] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.154831] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] cd8e2eaf-3c13-452b-b2e6-8107f0219378/cd8e2eaf-3c13-452b-b2e6-8107f0219378.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 933.154831] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f64163cb-265a-4a65-85a7-f31808ab279b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.161332] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 933.161332] env[61985]: value = "task-936137" [ 933.161332] env[61985]: _type = "Task" [ 933.161332] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.172079] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936137, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.173506] env[61985]: DEBUG nova.objects.instance [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lazy-loading 'flavor' on Instance uuid 1f284789-1e7b-4e9f-9670-34e8e25cd797 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.198661] env[61985]: DEBUG oslo_vmware.api [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936130, 'name': RemoveSnapshot_Task, 'duration_secs': 1.634005} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.198661] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 933.198661] env[61985]: INFO nova.compute.manager [None req-7b5fd9da-bcbc-4315-b473-961523217460 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Took 18.73 seconds to snapshot the instance on the hypervisor. [ 933.465737] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquiring lock "4da2d122-aa36-4cad-9521-498b53b9f2f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.466452] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lock "4da2d122-aa36-4cad-9521-498b53b9f2f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.504304] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936135, 'name': CreateVM_Task, 'duration_secs': 0.553991} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.504683] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 933.505622] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.506098] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.506507] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.506850] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40f36223-3f6b-48f8-8d4e-450f13477cdd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.515404] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 933.515404] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52815d1e-7156-0f96-13c2-360a73c5a3ec" [ 933.515404] env[61985]: _type = "Task" [ 933.515404] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.527094] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52815d1e-7156-0f96-13c2-360a73c5a3ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.575499] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936136, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.588501] env[61985]: ERROR nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [req-578c93f5-4b25-426d-b00b-8ee6f6ea51ce] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-578c93f5-4b25-426d-b00b-8ee6f6ea51ce"}]} [ 933.616567] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 933.628725] env[61985]: DEBUG nova.compute.manager [req-e2092b0d-9bb0-429e-985e-15b50f4236e9 req-91f44f99-d7d5-41dc-9b34-77d59c2761ba service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Received event network-changed-1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 933.630038] env[61985]: DEBUG nova.compute.manager [req-e2092b0d-9bb0-429e-985e-15b50f4236e9 req-91f44f99-d7d5-41dc-9b34-77d59c2761ba service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Refreshing instance network info cache due to event network-changed-1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 933.630038] env[61985]: DEBUG oslo_concurrency.lockutils [req-e2092b0d-9bb0-429e-985e-15b50f4236e9 req-91f44f99-d7d5-41dc-9b34-77d59c2761ba service nova] Acquiring lock "refresh_cache-488d0b95-97a9-4193-af62-cc92caf99625" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.630038] env[61985]: DEBUG oslo_concurrency.lockutils [req-e2092b0d-9bb0-429e-985e-15b50f4236e9 req-91f44f99-d7d5-41dc-9b34-77d59c2761ba service nova] Acquired lock "refresh_cache-488d0b95-97a9-4193-af62-cc92caf99625" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.630038] env[61985]: DEBUG nova.network.neutron [req-e2092b0d-9bb0-429e-985e-15b50f4236e9 req-91f44f99-d7d5-41dc-9b34-77d59c2761ba service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Refreshing network info cache for port 1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 933.638506] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 933.638506] env[61985]: DEBUG nova.compute.provider_tree [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 933.656132] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 933.673308] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936137, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.691076] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 933.971030] env[61985]: DEBUG nova.compute.manager [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.027969] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52815d1e-7156-0f96-13c2-360a73c5a3ec, 'name': SearchDatastore_Task, 'duration_secs': 0.05875} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.028341] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.028835] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.029059] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.029223] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.029443] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.032190] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41a07e9b-cbdc-419c-8f9f-2916fec1c2df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.049061] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.049351] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 934.050261] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeff76fc-60ff-44f3-9d79-9e17adce79bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.055511] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 934.055511] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524134f8-25d5-a40a-2db0-df0df262ce75" [ 934.055511] env[61985]: _type = "Task" [ 934.055511] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.066425] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524134f8-25d5-a40a-2db0-df0df262ce75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.074955] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936136, 'name': PowerOnVM_Task, 'duration_secs': 0.870412} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.078439] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 934.078439] env[61985]: INFO nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Took 9.24 seconds to spawn the instance on the hypervisor. [ 934.078577] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.079531] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08093544-31ef-4060-ac76-7409deee21dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.150340] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a0e213-fe37-4203-a8a9-d7571429069a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.158285] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b26a128-0882-4219-9221-549bfdfe494d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.190495] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4428c46c-3fcf-44fc-ab29-3c41ced04246 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.913s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.194037] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c53e97-c8e2-4941-a041-5ab8257b5e6f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.203375] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.203672] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.204867] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936137, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.643095} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.206101] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] cd8e2eaf-3c13-452b-b2e6-8107f0219378/cd8e2eaf-3c13-452b-b2e6-8107f0219378.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 934.206384] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.207136] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed463d62-0964-4a3b-8a0c-1aae8041535f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.215050] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebcc5923-c47e-44f4-b39c-8a8979e94f24 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.221161] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 934.221161] env[61985]: value = "task-936138" [ 934.221161] env[61985]: _type = "Task" [ 934.221161] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.234322] env[61985]: DEBUG nova.compute.provider_tree [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 934.242043] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936138, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.440870] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.441183] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.441406] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.441601] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.441834] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.444380] env[61985]: INFO nova.compute.manager [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Terminating instance [ 934.452140] env[61985]: DEBUG nova.compute.manager [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 934.452338] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 934.453499] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346b1575-1ec3-4fe5-972d-6420c0872a1f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.461079] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 934.461348] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41bef840-412b-4b94-a054-c030bbdbb749 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.471384] env[61985]: DEBUG oslo_vmware.api [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 934.471384] env[61985]: value = "task-936139" [ 934.471384] env[61985]: _type = "Task" [ 934.471384] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.482820] env[61985]: DEBUG oslo_vmware.api [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936139, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.500833] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9afb149a-6f1b-4848-9c28-8ae55b284243 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.504285] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.520639] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance '2322ab72-9841-41fb-9d60-2812baabe108' progress to 0 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 934.569023] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524134f8-25d5-a40a-2db0-df0df262ce75, 'name': SearchDatastore_Task, 'duration_secs': 0.021395} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.569402] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd6dba0f-9a5f-4541-8ec8-b1ba0ec34fb2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.575956] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 934.575956] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52964dab-53ee-73ac-5226-917513c92102" [ 934.575956] env[61985]: _type = "Task" [ 934.575956] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.589017] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52964dab-53ee-73ac-5226-917513c92102, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.599683] env[61985]: INFO nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Took 38.02 seconds to build instance. [ 934.648994] env[61985]: DEBUG nova.network.neutron [req-e2092b0d-9bb0-429e-985e-15b50f4236e9 req-91f44f99-d7d5-41dc-9b34-77d59c2761ba service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Updated VIF entry in instance network info cache for port 1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 934.649439] env[61985]: DEBUG nova.network.neutron [req-e2092b0d-9bb0-429e-985e-15b50f4236e9 req-91f44f99-d7d5-41dc-9b34-77d59c2761ba service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Updating instance_info_cache with network_info: [{"id": "1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4", "address": "fa:16:3e:b3:5d:0f", "network": {"id": "ddf63d11-2633-4be8-b73f-c5ca846b420c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-709607861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8015ef21eaf4a0bb19063e261064cdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bfe0ade-ba", "ovs_interfaceid": "1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.707438] env[61985]: DEBUG nova.compute.manager [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.735490] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936138, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10965} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.735850] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.737047] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f513fc8a-afdb-443e-a79d-200c6dda02df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.775542] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] cd8e2eaf-3c13-452b-b2e6-8107f0219378/cd8e2eaf-3c13-452b-b2e6-8107f0219378.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.777824] env[61985]: ERROR nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] [req-802b186a-c979-48c2-b37b-b96ecfb1d6d4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-802b186a-c979-48c2-b37b-b96ecfb1d6d4"}]} [ 934.778423] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d122e5e0-aa11-437c-8c01-f71624200b21 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.809690] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 934.809690] env[61985]: value = "task-936140" [ 934.809690] env[61985]: _type = "Task" [ 934.809690] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.820936] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936140, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.822293] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 934.840615] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 934.840743] env[61985]: DEBUG nova.compute.provider_tree [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 934.854481] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 934.878439] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 934.983975] env[61985]: DEBUG oslo_vmware.api [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936139, 'name': PowerOffVM_Task, 'duration_secs': 0.283435} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.984301] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 934.984488] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 934.984758] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7be149be-c4a7-4251-a931-620055e9f69d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.034083] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 935.034452] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58d33a21-3c7d-47e1-a384-553418978302 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.047430] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 935.047430] env[61985]: value = "task-936142" [ 935.047430] env[61985]: _type = "Task" [ 935.047430] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.056785] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 935.056785] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 935.056785] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleting the datastore file [datastore1] 5c2a7072-5b14-4ec7-8060-4e21eea927e3 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.056785] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7174f9bc-32cf-4f42-892b-408fed3f0512 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.063370] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 935.063370] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance '2322ab72-9841-41fb-9d60-2812baabe108' progress to 17 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 935.070718] env[61985]: DEBUG oslo_vmware.api [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 935.070718] env[61985]: value = "task-936143" [ 935.070718] env[61985]: _type = "Task" [ 935.070718] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.083615] env[61985]: DEBUG oslo_vmware.api [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936143, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.091182] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52964dab-53ee-73ac-5226-917513c92102, 'name': SearchDatastore_Task, 'duration_secs': 0.021472} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.091743] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.093581] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 488d0b95-97a9-4193-af62-cc92caf99625/488d0b95-97a9-4193-af62-cc92caf99625.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 935.093581] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d719c99f-3bf9-451e-a836-eddbb75b83c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.099952] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 935.099952] env[61985]: value = "task-936144" [ 935.099952] env[61985]: _type = "Task" [ 935.099952] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.100451] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "62824eac-9412-466a-abcf-1010f6a829e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.533s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.111956] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936144, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.153036] env[61985]: DEBUG oslo_concurrency.lockutils [req-e2092b0d-9bb0-429e-985e-15b50f4236e9 req-91f44f99-d7d5-41dc-9b34-77d59c2761ba service nova] Releasing lock "refresh_cache-488d0b95-97a9-4193-af62-cc92caf99625" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.237693] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.309561] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "6dd8218b-c14d-40c9-87df-097fab06c669" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.309947] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.329519] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936140, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.331868] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342b0621-44f1-4672-acdf-a973e5ee66df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.338578] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.338832] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.341834] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8b661d-d91e-4ea7-8499-d43715d68d97 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.375721] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f63fdfe-4fcf-477b-8874-20d046f06d6d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.385105] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83d0fcd-0cf6-4343-bb88-790b60ea3260 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.399820] env[61985]: DEBUG nova.compute.provider_tree [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.548828] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquiring lock "518635bf-73ee-404b-ae6a-dc4ee23009d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.549115] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lock "518635bf-73ee-404b-ae6a-dc4ee23009d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.549343] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquiring lock "518635bf-73ee-404b-ae6a-dc4ee23009d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.549576] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lock "518635bf-73ee-404b-ae6a-dc4ee23009d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.549714] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lock "518635bf-73ee-404b-ae6a-dc4ee23009d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.552119] env[61985]: INFO nova.compute.manager [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Terminating instance [ 935.554257] env[61985]: DEBUG nova.compute.manager [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 935.554479] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 935.555350] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628b241a-05d5-4b8c-a0c2-32fb362c42ec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.564229] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 935.564524] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d15137d2-3586-4d18-a506-d006993e1c52 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.575810] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.576321] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.576568] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.576781] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.576934] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.577118] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.577346] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.577515] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.577833] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.578022] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.578253] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.585496] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ce91119-8301-419f-a637-abe2f2b0a13c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.601734] env[61985]: DEBUG oslo_vmware.api [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 935.601734] env[61985]: value = "task-936145" [ 935.601734] env[61985]: _type = "Task" [ 935.601734] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.613786] env[61985]: DEBUG oslo_vmware.api [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936143, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.307576} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.614489] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 935.614489] env[61985]: value = "task-936146" [ 935.614489] env[61985]: _type = "Task" [ 935.614489] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.615434] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.615434] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 935.615434] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 935.615654] env[61985]: INFO nova.compute.manager [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Took 1.16 seconds to destroy the instance on the hypervisor. [ 935.615879] env[61985]: DEBUG oslo.service.loopingcall [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.616205] env[61985]: DEBUG nova.compute.manager [-] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.616303] env[61985]: DEBUG nova.network.neutron [-] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 935.630714] env[61985]: DEBUG oslo_vmware.api [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936145, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.631010] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936144, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.637106] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936146, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.813948] env[61985]: DEBUG nova.compute.utils [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 935.829683] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936140, 'name': ReconfigVM_Task, 'duration_secs': 0.616064} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.830182] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Reconfigured VM instance instance-0000004b to attach disk [datastore2] cd8e2eaf-3c13-452b-b2e6-8107f0219378/cd8e2eaf-3c13-452b-b2e6-8107f0219378.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.831084] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ebcd7d3e-d913-4697-a5cc-8b9405b761c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.838624] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 935.838624] env[61985]: value = "task-936147" [ 935.838624] env[61985]: _type = "Task" [ 935.838624] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.843048] env[61985]: INFO nova.compute.manager [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Detaching volume 78b9d30c-3370-4505-b9bf-60bae00d9eda [ 935.850782] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936147, 'name': Rename_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.887367] env[61985]: INFO nova.virt.block_device [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Attempting to driver detach volume 78b9d30c-3370-4505-b9bf-60bae00d9eda from mountpoint /dev/sdc [ 935.887736] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 935.887977] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211408', 'volume_id': '78b9d30c-3370-4505-b9bf-60bae00d9eda', 'name': 'volume-78b9d30c-3370-4505-b9bf-60bae00d9eda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1f284789-1e7b-4e9f-9670-34e8e25cd797', 'attached_at': '', 'detached_at': '', 'volume_id': '78b9d30c-3370-4505-b9bf-60bae00d9eda', 'serial': '78b9d30c-3370-4505-b9bf-60bae00d9eda'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 935.888956] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369c65f1-8eab-4ae2-aa33-90ab2f9c216f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.922283] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ff83ff-289c-4395-99a7-e39fd0bed1c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.932088] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3509bdca-622a-418f-a0b8-82a997f628a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.958512] env[61985]: DEBUG nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 99 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 935.958804] env[61985]: DEBUG nova.compute.provider_tree [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 99 to 100 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 935.959184] env[61985]: DEBUG nova.compute.provider_tree [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.963741] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b00529-78f1-4f5f-9e84-b6f702f6636e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.982834] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] The volume has not been displaced from its original location: [datastore1] volume-78b9d30c-3370-4505-b9bf-60bae00d9eda/volume-78b9d30c-3370-4505-b9bf-60bae00d9eda.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 935.988511] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Reconfiguring VM instance instance-0000003a to detach disk 2002 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 935.990775] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-663f4bc2-b9bb-44ce-97d9-3e3d8d248013 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.005151] env[61985]: DEBUG nova.network.neutron [-] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.011625] env[61985]: DEBUG oslo_vmware.api [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 936.011625] env[61985]: value = "task-936148" [ 936.011625] env[61985]: _type = "Task" [ 936.011625] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.020133] env[61985]: DEBUG oslo_vmware.api [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936148, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.120067] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936144, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659942} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.120405] env[61985]: DEBUG oslo_vmware.api [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936145, 'name': PowerOffVM_Task, 'duration_secs': 0.396028} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.123438] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 488d0b95-97a9-4193-af62-cc92caf99625/488d0b95-97a9-4193-af62-cc92caf99625.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 936.123680] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.123931] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 936.124111] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 936.124347] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e9639ea-6923-4dbb-873c-55fb3d5eb8ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.126552] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4ebfc03-3084-4a8e-97fa-e76042045321 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.132362] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936146, 'name': ReconfigVM_Task, 'duration_secs': 0.183509} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.133784] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance '2322ab72-9841-41fb-9d60-2812baabe108' progress to 33 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 936.137922] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 936.137922] env[61985]: value = "task-936149" [ 936.137922] env[61985]: _type = "Task" [ 936.137922] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.148692] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936149, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.194763] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 936.195035] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 936.195231] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Deleting the datastore file [datastore2] 518635bf-73ee-404b-ae6a-dc4ee23009d2 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.195513] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9aa74748-9a10-466d-a1ad-eb974e4cf529 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.202838] env[61985]: DEBUG oslo_vmware.api [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for the task: (returnval){ [ 936.202838] env[61985]: value = "task-936151" [ 936.202838] env[61985]: _type = "Task" [ 936.202838] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.210848] env[61985]: DEBUG oslo_vmware.api [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.309429] env[61985]: DEBUG nova.compute.manager [req-a0962276-a651-40d1-b7cb-b6350913e0e4 req-2629bc0f-b33c-46fb-9b9d-2d2524b08107 service nova] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Received event network-vif-deleted-41b487bd-afda-4a0c-be1a-056e516b010c {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 936.323486] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.014s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.350309] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936147, 'name': Rename_Task, 'duration_secs': 0.166612} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.350680] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 936.350956] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-21ec57da-3f9d-4369-9d47-94a36c461686 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.357608] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 936.357608] env[61985]: value = "task-936152" [ 936.357608] env[61985]: _type = "Task" [ 936.357608] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.365744] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.469609] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 5.913s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.473243] env[61985]: DEBUG oslo_concurrency.lockutils [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.967s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.473243] env[61985]: DEBUG nova.objects.instance [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61985) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 936.495100] env[61985]: INFO nova.scheduler.client.report [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Deleted allocations for instance 54ba5cfd-185a-4c58-aa5e-83cc096a482e [ 936.509831] env[61985]: INFO nova.compute.manager [-] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Took 0.89 seconds to deallocate network for instance. [ 936.527314] env[61985]: DEBUG oslo_vmware.api [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936148, 'name': ReconfigVM_Task, 'duration_secs': 0.401647} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.527652] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Reconfigured VM instance instance-0000003a to detach disk 2002 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 936.533157] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43b255c3-8234-42ee-8427-93dead98766f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.550731] env[61985]: DEBUG oslo_vmware.api [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 936.550731] env[61985]: value = "task-936153" [ 936.550731] env[61985]: _type = "Task" [ 936.550731] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.562399] env[61985]: DEBUG oslo_vmware.api [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936153, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.640227] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.640556] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.640719] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.640921] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.641091] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.641302] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.641547] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.641745] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.641927] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.642340] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.642407] env[61985]: DEBUG nova.virt.hardware [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.648121] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Reconfiguring VM instance instance-00000025 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 936.648519] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3a937d1-b22e-4a4b-8a1e-8204c68e61b2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.673656] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936149, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068889} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.678022] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.678022] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 936.678022] env[61985]: value = "task-936154" [ 936.678022] env[61985]: _type = "Task" [ 936.678022] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.678022] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a565f414-7496-480b-8a1b-f87459c23f8f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.704075] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 488d0b95-97a9-4193-af62-cc92caf99625/488d0b95-97a9-4193-af62-cc92caf99625.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.708273] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4983ff81-8428-49f7-9489-c35f07e6cad2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.724942] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.735718] env[61985]: DEBUG oslo_vmware.api [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Task: {'id': task-936151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.329299} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.737118] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 936.737360] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 936.737564] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 936.737764] env[61985]: INFO nova.compute.manager [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Took 1.18 seconds to destroy the instance on the hypervisor. [ 936.738058] env[61985]: DEBUG oslo.service.loopingcall [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.738355] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 936.738355] env[61985]: value = "task-936155" [ 936.738355] env[61985]: _type = "Task" [ 936.738355] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.738556] env[61985]: DEBUG nova.compute.manager [-] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 936.738675] env[61985]: DEBUG nova.network.neutron [-] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 936.749572] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936155, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.868689] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936152, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.005642] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06bafcdb-b28b-435b-9497-cb4208c1f750 tempest-ServerAddressesNegativeTestJSON-1682828482 tempest-ServerAddressesNegativeTestJSON-1682828482-project-member] Lock "54ba5cfd-185a-4c58-aa5e-83cc096a482e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.620s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.016903] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.061682] env[61985]: DEBUG oslo_vmware.api [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936153, 'name': ReconfigVM_Task, 'duration_secs': 0.258845} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.062009] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211408', 'volume_id': '78b9d30c-3370-4505-b9bf-60bae00d9eda', 'name': 'volume-78b9d30c-3370-4505-b9bf-60bae00d9eda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1f284789-1e7b-4e9f-9670-34e8e25cd797', 'attached_at': '', 'detached_at': '', 'volume_id': '78b9d30c-3370-4505-b9bf-60bae00d9eda', 'serial': '78b9d30c-3370-4505-b9bf-60bae00d9eda'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 937.190503] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936154, 'name': ReconfigVM_Task, 'duration_secs': 0.22391} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.190823] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Reconfigured VM instance instance-00000025 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 937.191697] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10a0715-bb6f-43fe-ac5e-e672c8517c53 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.215800] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 2322ab72-9841-41fb-9d60-2812baabe108/2322ab72-9841-41fb-9d60-2812baabe108.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.216155] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec18c8af-2ed9-4e18-a8f6-13ce970d4061 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.234676] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 937.234676] env[61985]: value = "task-936156" [ 937.234676] env[61985]: _type = "Task" [ 937.234676] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.242951] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936156, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.252092] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936155, 'name': ReconfigVM_Task, 'duration_secs': 0.318474} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.252268] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 488d0b95-97a9-4193-af62-cc92caf99625/488d0b95-97a9-4193-af62-cc92caf99625.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.253026] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fad85b16-5345-4d93-8f69-a772fdcda438 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.260232] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 937.260232] env[61985]: value = "task-936157" [ 937.260232] env[61985]: _type = "Task" [ 937.260232] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.268790] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936157, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.368591] env[61985]: DEBUG oslo_vmware.api [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936152, 'name': PowerOnVM_Task, 'duration_secs': 0.654363} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.368860] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 937.369083] env[61985]: INFO nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Took 9.95 seconds to spawn the instance on the hypervisor. [ 937.369357] env[61985]: DEBUG nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.370237] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f725bd-5389-4069-97b9-1c4f0da267cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.395775] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "6dd8218b-c14d-40c9-87df-097fab06c669" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.397795] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.397795] env[61985]: INFO nova.compute.manager [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Attaching volume 7edc010c-38ef-42c0-af8a-550bed9ea645 to /dev/sdb [ 937.435542] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e146de-ccde-46eb-85ce-625ec91704aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.445814] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b067510-8e86-43bd-8fe3-32be4c20e224 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.460270] env[61985]: DEBUG nova.virt.block_device [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updating existing volume attachment record: 188f7f4d-824b-4eb7-af11-184156d515d9 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 937.483472] env[61985]: DEBUG oslo_concurrency.lockutils [None req-db4dd4cc-77b5-481b-9d2f-000f84bdf88e tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.484694] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.825s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.490662] env[61985]: INFO nova.compute.claims [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.610832] env[61985]: DEBUG nova.objects.instance [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lazy-loading 'flavor' on Instance uuid 1f284789-1e7b-4e9f-9670-34e8e25cd797 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.612374] env[61985]: DEBUG nova.network.neutron [-] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.745141] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936156, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.771591] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936157, 'name': Rename_Task, 'duration_secs': 0.135069} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.772061] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 937.772360] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e3520fd-3009-4dd4-acf5-e9e8e7f59386 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.778779] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 937.778779] env[61985]: value = "task-936160" [ 937.778779] env[61985]: _type = "Task" [ 937.778779] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.787754] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936160, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.890207] env[61985]: INFO nova.compute.manager [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Took 41.26 seconds to build instance. [ 938.116357] env[61985]: INFO nova.compute.manager [-] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Took 1.38 seconds to deallocate network for instance. [ 938.249516] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936156, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.293362] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936160, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.392943] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2c43319e-30af-4e01-9562-086fa6773114 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.774s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.556777] env[61985]: DEBUG nova.compute.manager [req-d411525a-3c16-4823-84f5-59c91a139b87 req-59b5e5d1-71c1-4cb8-84a5-46fd6e0e305a service nova] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Received event network-vif-deleted-cfee741d-7a1d-4e4f-8432-1603248a1ba4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 938.603127] env[61985]: DEBUG oslo_concurrency.lockutils [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.603785] env[61985]: DEBUG oslo_concurrency.lockutils [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.603910] env[61985]: DEBUG oslo_concurrency.lockutils [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.604133] env[61985]: DEBUG oslo_concurrency.lockutils [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.604626] env[61985]: DEBUG oslo_concurrency.lockutils [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.609683] env[61985]: INFO nova.compute.manager [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Terminating instance [ 938.612467] env[61985]: DEBUG nova.compute.manager [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 938.612467] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 938.613042] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b36977-ffb7-497a-b39d-36e174320f0e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.622019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3c02c8e6-9350-47ea-a15b-1e35b6b6109e tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.282s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.626324] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.626779] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 938.629809] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cd033ecb-e742-4be0-bec7-932612c8844e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.641093] env[61985]: DEBUG oslo_vmware.api [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 938.641093] env[61985]: value = "task-936162" [ 938.641093] env[61985]: _type = "Task" [ 938.641093] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.650921] env[61985]: DEBUG oslo_vmware.api [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936162, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.747104] env[61985]: DEBUG oslo_vmware.api [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936156, 'name': ReconfigVM_Task, 'duration_secs': 1.043346} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.749772] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 2322ab72-9841-41fb-9d60-2812baabe108/2322ab72-9841-41fb-9d60-2812baabe108.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.750180] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance '2322ab72-9841-41fb-9d60-2812baabe108' progress to 50 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 938.774545] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "62824eac-9412-466a-abcf-1010f6a829e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.775035] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "62824eac-9412-466a-abcf-1010f6a829e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.775350] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "62824eac-9412-466a-abcf-1010f6a829e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.775605] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "62824eac-9412-466a-abcf-1010f6a829e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.775803] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "62824eac-9412-466a-abcf-1010f6a829e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.779800] env[61985]: INFO nova.compute.manager [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Terminating instance [ 938.789905] env[61985]: DEBUG nova.compute.manager [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 938.789905] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 938.789905] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe95b0f-2b68-43a8-b1f4-523337c8c821 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.801474] env[61985]: DEBUG oslo_vmware.api [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936160, 'name': PowerOnVM_Task, 'duration_secs': 0.530091} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.805037] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 938.805037] env[61985]: INFO nova.compute.manager [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Took 8.80 seconds to spawn the instance on the hypervisor. [ 938.805037] env[61985]: DEBUG nova.compute.manager [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.805037] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 938.808154] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f937bf-617f-4938-b8c7-07953e5c8748 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.808154] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ea572a8-379e-4e85-8fe6-1f09b65e27b1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.817436] env[61985]: DEBUG oslo_vmware.api [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 938.817436] env[61985]: value = "task-936163" [ 938.817436] env[61985]: _type = "Task" [ 938.817436] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.827819] env[61985]: DEBUG oslo_vmware.api [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936163, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.009623] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b670255f-93b5-4e56-ae7b-6cd80eade3d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.018202] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90fb4144-5206-4e33-bb69-9c1d2edc1d18 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.055711] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aead7811-db9f-4cb5-8ec2-da65de00431e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.063760] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8dbbaf-7bf0-48cd-a771-a4ac3b2a3a2e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.078542] env[61985]: DEBUG nova.compute.provider_tree [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.133484] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.133929] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.133929] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "1f284789-1e7b-4e9f-9670-34e8e25cd797-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.134150] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.134363] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.136813] env[61985]: INFO nova.compute.manager [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Terminating instance [ 939.139042] env[61985]: DEBUG nova.compute.manager [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 939.139244] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 939.140132] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190c21db-b040-4096-8eb8-dcf72880f5b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.154955] env[61985]: DEBUG oslo_vmware.api [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936162, 'name': PowerOffVM_Task, 'duration_secs': 0.249876} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.155650] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 939.155788] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 939.155963] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 939.156201] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d0a3c82-45e9-4c5b-b22a-a498077a826f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.157713] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3dcc781a-92e1-406f-8f17-76388e99ab9a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.164442] env[61985]: DEBUG oslo_vmware.api [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 939.164442] env[61985]: value = "task-936164" [ 939.164442] env[61985]: _type = "Task" [ 939.164442] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.173200] env[61985]: DEBUG oslo_vmware.api [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936164, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.227025] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 939.227274] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 939.227468] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleting the datastore file [datastore2] cd8e2eaf-3c13-452b-b2e6-8107f0219378 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.227818] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc9a0a4b-01bd-4eca-a78b-4baf19b336b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.234626] env[61985]: DEBUG oslo_vmware.api [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 939.234626] env[61985]: value = "task-936166" [ 939.234626] env[61985]: _type = "Task" [ 939.234626] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.243193] env[61985]: DEBUG oslo_vmware.api [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936166, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.256847] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960c4a33-2fdc-4585-93f8-78fe00c0ee35 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.276347] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05eaa8e-9fcb-4469-8a08-16598680d0b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.295379] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance '2322ab72-9841-41fb-9d60-2812baabe108' progress to 67 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 939.328902] env[61985]: INFO nova.compute.manager [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Took 39.60 seconds to build instance. [ 939.332601] env[61985]: DEBUG oslo_vmware.api [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936163, 'name': PowerOffVM_Task, 'duration_secs': 0.214818} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.333506] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 939.333506] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 939.333506] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f26f104e-3654-40a4-904f-42a0a9904c4a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.478534] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 939.478610] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 939.478779] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleting the datastore file [datastore2] 62824eac-9412-466a-abcf-1010f6a829e4 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.479064] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5202efd3-674d-4e4a-9d68-60a07ff6616d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.485840] env[61985]: DEBUG oslo_vmware.api [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for the task: (returnval){ [ 939.485840] env[61985]: value = "task-936168" [ 939.485840] env[61985]: _type = "Task" [ 939.485840] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.494022] env[61985]: DEBUG oslo_vmware.api [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936168, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.583369] env[61985]: DEBUG nova.scheduler.client.report [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.676173] env[61985]: DEBUG oslo_vmware.api [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936164, 'name': PowerOffVM_Task, 'duration_secs': 0.272137} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.676568] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 939.676708] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 939.677112] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dfe37f51-b74d-42c3-b1c3-5515a45d954d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.745173] env[61985]: DEBUG oslo_vmware.api [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936166, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.340167} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.746369] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.746673] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 939.746973] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 939.747292] env[61985]: INFO nova.compute.manager [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Took 1.14 seconds to destroy the instance on the hypervisor. [ 939.747732] env[61985]: DEBUG oslo.service.loopingcall [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.748068] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 939.748281] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 939.748462] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Deleting the datastore file [datastore2] 1f284789-1e7b-4e9f-9670-34e8e25cd797 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.748710] env[61985]: DEBUG nova.compute.manager [-] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.748823] env[61985]: DEBUG nova.network.neutron [-] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 939.750568] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95b2a275-4234-485d-b8db-52559ae74ca1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.756745] env[61985]: DEBUG oslo_vmware.api [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for the task: (returnval){ [ 939.756745] env[61985]: value = "task-936170" [ 939.756745] env[61985]: _type = "Task" [ 939.756745] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.764726] env[61985]: DEBUG oslo_vmware.api [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.833882] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f3b93d4a-233a-4fb4-a294-2d044003b637 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "488d0b95-97a9-4193-af62-cc92caf99625" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.125s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.879400] env[61985]: DEBUG nova.network.neutron [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Port 4aa18b7b-89b5-4cad-af09-1bbc8f255029 binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 939.995885] env[61985]: DEBUG oslo_vmware.api [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Task: {'id': task-936168, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228554} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.996185] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.996394] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 939.996587] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 939.996768] env[61985]: INFO nova.compute.manager [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Took 1.21 seconds to destroy the instance on the hypervisor. [ 939.997123] env[61985]: DEBUG oslo.service.loopingcall [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.997239] env[61985]: DEBUG nova.compute.manager [-] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.997335] env[61985]: DEBUG nova.network.neutron [-] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 940.090113] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.090683] env[61985]: DEBUG nova.compute.manager [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 940.094031] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.030s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.094357] env[61985]: DEBUG nova.objects.instance [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lazy-loading 'resources' on Instance uuid e68cc5ff-83b7-4602-be50-ee4b1fa9ac65 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.268989] env[61985]: DEBUG oslo_vmware.api [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Task: {'id': task-936170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.2079} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.269357] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.269477] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 940.269687] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 940.269927] env[61985]: INFO nova.compute.manager [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Took 1.13 seconds to destroy the instance on the hypervisor. [ 940.270324] env[61985]: DEBUG oslo.service.loopingcall [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.270491] env[61985]: DEBUG nova.compute.manager [-] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.270557] env[61985]: DEBUG nova.network.neutron [-] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 940.556588] env[61985]: DEBUG nova.network.neutron [-] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.601676] env[61985]: DEBUG nova.compute.utils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.607755] env[61985]: DEBUG nova.compute.manager [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 940.607989] env[61985]: DEBUG nova.network.neutron [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 940.903638] env[61985]: DEBUG nova.policy [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '344f7f484f4240cd9dc32b200b697e12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfc57e6393c94545be32165d41230db3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.927702] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "2322ab72-9841-41fb-9d60-2812baabe108-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.928968] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.928968] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.933572] env[61985]: DEBUG nova.network.neutron [-] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.954786] env[61985]: DEBUG nova.compute.manager [req-d56f1fc2-88be-4219-9368-698670663539 req-26019c4c-8458-43b7-9e3c-8e2ebad40905 service nova] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Received event network-vif-deleted-57bd4c36-e5e2-43f4-b8e5-7b574b97ba90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 940.955607] env[61985]: DEBUG nova.compute.manager [req-d56f1fc2-88be-4219-9368-698670663539 req-26019c4c-8458-43b7-9e3c-8e2ebad40905 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Received event network-vif-deleted-6ee77311-001c-4c2f-afad-e75094a76022 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 940.956053] env[61985]: INFO nova.compute.manager [req-d56f1fc2-88be-4219-9368-698670663539 req-26019c4c-8458-43b7-9e3c-8e2ebad40905 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Neutron deleted interface 6ee77311-001c-4c2f-afad-e75094a76022; detaching it from the instance and deleting it from the info cache [ 940.956292] env[61985]: DEBUG nova.network.neutron [req-d56f1fc2-88be-4219-9368-698670663539 req-26019c4c-8458-43b7-9e3c-8e2ebad40905 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.060799] env[61985]: INFO nova.compute.manager [-] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Took 1.31 seconds to deallocate network for instance. [ 941.083465] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-765a27e2-41e3-45f3-9cf7-cb395ef6f18d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.092143] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c5a8bf-1f57-404a-9a51-522a70eda997 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.130949] env[61985]: DEBUG nova.compute.manager [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 941.135331] env[61985]: DEBUG nova.compute.manager [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.136241] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57206c2-afda-48da-b35e-583d5c702f82 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.142359] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a6c6e1-35e6-4bc2-9b26-3fcd1166cb60 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.157396] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6668561b-0c49-4b75-8557-7fca1a96d722 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.172831] env[61985]: DEBUG nova.compute.provider_tree [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.435956] env[61985]: INFO nova.compute.manager [-] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Took 1.44 seconds to deallocate network for instance. [ 941.464292] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9913e8a9-77b0-4850-98d8-cbccdc67d95f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.477063] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c912b76-68f3-41bd-a6f7-a6498e51d5b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.514416] env[61985]: DEBUG nova.compute.manager [req-d56f1fc2-88be-4219-9368-698670663539 req-26019c4c-8458-43b7-9e3c-8e2ebad40905 service nova] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Detach interface failed, port_id=6ee77311-001c-4c2f-afad-e75094a76022, reason: Instance 62824eac-9412-466a-abcf-1010f6a829e4 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 941.570113] env[61985]: DEBUG oslo_concurrency.lockutils [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.622976] env[61985]: DEBUG nova.network.neutron [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Successfully created port: 262813ed-9baf-4240-b57d-fc4a648b5532 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.658402] env[61985]: INFO nova.compute.manager [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] instance snapshotting [ 941.660708] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7786b6-d36d-42bd-ad50-394f03ead12b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.683933] env[61985]: DEBUG nova.scheduler.client.report [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.693128] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa978829-8e50-487b-a380-3b3f450f0eb0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.950825] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.972028] env[61985]: DEBUG nova.network.neutron [-] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.014879] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 942.015140] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211422', 'volume_id': '7edc010c-38ef-42c0-af8a-550bed9ea645', 'name': 'volume-7edc010c-38ef-42c0-af8a-550bed9ea645', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dd8218b-c14d-40c9-87df-097fab06c669', 'attached_at': '', 'detached_at': '', 'volume_id': '7edc010c-38ef-42c0-af8a-550bed9ea645', 'serial': '7edc010c-38ef-42c0-af8a-550bed9ea645'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 942.016101] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a093c194-886f-4b31-935d-24771f6afb47 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.039795] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312e7e49-12a7-4129-9c74-43da9ded605c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.070850] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] volume-7edc010c-38ef-42c0-af8a-550bed9ea645/volume-7edc010c-38ef-42c0-af8a-550bed9ea645.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.072026] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.072212] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.072393] env[61985]: DEBUG nova.network.neutron [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 942.074595] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7512e60c-4453-407c-b761-59d904a35a1d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.095723] env[61985]: DEBUG oslo_vmware.api [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 942.095723] env[61985]: value = "task-936172" [ 942.095723] env[61985]: _type = "Task" [ 942.095723] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.106942] env[61985]: DEBUG oslo_vmware.api [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936172, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.145692] env[61985]: DEBUG nova.compute.manager [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 942.177974] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.179137] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.179137] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.179137] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.179137] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.179137] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.179580] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.179626] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.179814] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.179989] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.180283] env[61985]: DEBUG nova.virt.hardware [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.181497] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8190a26a-d080-40f0-ab13-37c164fd27e5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.192039] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a3bc7d-0868-4e39-abf4-6610bac4d8c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.199050] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.105s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.214164] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.433s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.214896] env[61985]: DEBUG nova.objects.instance [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'resources' on Instance uuid 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.217138] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 942.218108] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e3ac65a7-e0cc-4839-933d-6d5c2868c610 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.226708] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 942.226708] env[61985]: value = "task-936173" [ 942.226708] env[61985]: _type = "Task" [ 942.226708] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.237870] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936173, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.239255] env[61985]: INFO nova.scheduler.client.report [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted allocations for instance e68cc5ff-83b7-4602-be50-ee4b1fa9ac65 [ 942.471812] env[61985]: INFO nova.compute.manager [-] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Took 2.20 seconds to deallocate network for instance. [ 942.607653] env[61985]: DEBUG oslo_vmware.api [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936172, 'name': ReconfigVM_Task, 'duration_secs': 0.464831} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.608038] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Reconfigured VM instance instance-00000043 to attach disk [datastore2] volume-7edc010c-38ef-42c0-af8a-550bed9ea645/volume-7edc010c-38ef-42c0-af8a-550bed9ea645.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.613695] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33cbbaf9-89d5-437a-a99c-8d9f173bf00d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.629741] env[61985]: DEBUG oslo_vmware.api [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 942.629741] env[61985]: value = "task-936174" [ 942.629741] env[61985]: _type = "Task" [ 942.629741] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.641020] env[61985]: DEBUG oslo_vmware.api [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936174, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.740853] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936173, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.752498] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d11f4f88-c27c-4c07-bd36-6ef457755945 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "e68cc5ff-83b7-4602-be50-ee4b1fa9ac65" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.711s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.981159] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.984955] env[61985]: DEBUG nova.compute.manager [req-2d5979d0-9f8d-48b3-99a0-f3d546e59052 req-6b141df3-fecd-44d6-87a1-f649187a057a service nova] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Received event network-vif-deleted-c4da38a8-666e-4d80-8841-66ef028ba74a {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 943.004359] env[61985]: DEBUG nova.network.neutron [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance_info_cache with network_info: [{"id": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "address": "fa:16:3e:12:45:ef", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aa18b7b-89", "ovs_interfaceid": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.118484] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d02100-ec32-414b-a04a-0ff36437510b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.129122] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b16725-6d4d-4e3d-a575-8b0e3a687634 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.167222] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26810eb5-6ba5-462e-9fd0-7c3d334e985a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.169794] env[61985]: DEBUG oslo_vmware.api [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936174, 'name': ReconfigVM_Task, 'duration_secs': 0.145591} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.170099] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211422', 'volume_id': '7edc010c-38ef-42c0-af8a-550bed9ea645', 'name': 'volume-7edc010c-38ef-42c0-af8a-550bed9ea645', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dd8218b-c14d-40c9-87df-097fab06c669', 'attached_at': '', 'detached_at': '', 'volume_id': '7edc010c-38ef-42c0-af8a-550bed9ea645', 'serial': '7edc010c-38ef-42c0-af8a-550bed9ea645'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 943.177033] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6a1170-ea6f-4e68-b1c0-d86db1e8836a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.191750] env[61985]: DEBUG nova.compute.provider_tree [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.223780] env[61985]: DEBUG oslo_vmware.rw_handles [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d3c40a-c43f-6b3d-a50b-2b96eb6c2649/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 943.224781] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fb1183-010c-49d4-90db-420b7abbd5c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.237784] env[61985]: DEBUG oslo_vmware.rw_handles [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d3c40a-c43f-6b3d-a50b-2b96eb6c2649/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 943.237979] env[61985]: ERROR oslo_vmware.rw_handles [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d3c40a-c43f-6b3d-a50b-2b96eb6c2649/disk-0.vmdk due to incomplete transfer. [ 943.241142] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6f5e0003-aad4-4247-8213-961446d801b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.242658] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936173, 'name': CreateSnapshot_Task, 'duration_secs': 0.773669} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.242911] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 943.245057] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519469eb-474c-4aed-acd7-da18e5705069 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.249100] env[61985]: DEBUG oslo_vmware.rw_handles [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d3c40a-c43f-6b3d-a50b-2b96eb6c2649/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 943.249973] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Uploaded image b4b4e229-a7cf-45de-8912-47f4b206140b to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 943.252749] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 943.256455] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a654e172-07dc-4eb1-a8ca-b4492232b324 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.268131] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 943.268131] env[61985]: value = "task-936175" [ 943.268131] env[61985]: _type = "Task" [ 943.268131] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.277648] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936175, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.479409] env[61985]: DEBUG nova.network.neutron [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Successfully updated port: 262813ed-9baf-4240-b57d-fc4a648b5532 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.509326] env[61985]: DEBUG oslo_concurrency.lockutils [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.694754] env[61985]: DEBUG nova.scheduler.client.report [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.771229] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 943.771626] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f5d9820e-38e5-4c7d-90f2-ecc50511222b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.783951] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936175, 'name': Destroy_Task, 'duration_secs': 0.297785} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.785305] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Destroyed the VM [ 943.785591] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 943.785933] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 943.785933] env[61985]: value = "task-936176" [ 943.785933] env[61985]: _type = "Task" [ 943.785933] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.786146] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e608e973-b021-4553-9b08-351bdea1e558 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.796032] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936176, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.797173] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 943.797173] env[61985]: value = "task-936177" [ 943.797173] env[61985]: _type = "Task" [ 943.797173] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.804603] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936177, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.981904] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.982087] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.982234] env[61985]: DEBUG nova.network.neutron [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 944.040881] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6c53da-22ed-4bdd-86d1-628f517fa896 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.068869] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47258c4-e7ad-4d18-a308-bbf0c2b2efc0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.078796] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance '2322ab72-9841-41fb-9d60-2812baabe108' progress to 83 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 944.199603] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.985s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.204161] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 13.861s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.207696] env[61985]: DEBUG nova.objects.instance [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lazy-loading 'flavor' on Instance uuid 6dd8218b-c14d-40c9-87df-097fab06c669 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.220077] env[61985]: INFO nova.scheduler.client.report [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Deleted allocations for instance 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf [ 944.298478] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936176, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.306331] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936177, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.530396] env[61985]: DEBUG nova.network.neutron [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 944.587136] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-63d3c3f4-8f9e-432a-bf4f-22d87f59e7a3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance '2322ab72-9841-41fb-9d60-2812baabe108' progress to 100 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 944.710385] env[61985]: INFO nova.compute.claims [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.718687] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41294c61-d6ea-489b-b48e-e7fb52c3e37e tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.322s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.730833] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a9782755-a847-4632-845e-694df77870ca tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.139s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.784619] env[61985]: DEBUG nova.network.neutron [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance_info_cache with network_info: [{"id": "262813ed-9baf-4240-b57d-fc4a648b5532", "address": "fa:16:3e:d3:c1:6a", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262813ed-9b", "ovs_interfaceid": "262813ed-9baf-4240-b57d-fc4a648b5532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.799622] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936176, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.808844] env[61985]: DEBUG oslo_vmware.api [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936177, 'name': RemoveSnapshot_Task, 'duration_secs': 0.602106} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.809193] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 944.809371] env[61985]: INFO nova.compute.manager [None req-eeae7edb-3ecd-4721-b467-d81aa5bbae48 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Took 17.24 seconds to snapshot the instance on the hypervisor. [ 944.890306] env[61985]: INFO nova.compute.manager [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Rescuing [ 944.890592] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.890756] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.890936] env[61985]: DEBUG nova.network.neutron [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 945.149886] env[61985]: DEBUG nova.compute.manager [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Received event network-vif-plugged-262813ed-9baf-4240-b57d-fc4a648b5532 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 945.154015] env[61985]: DEBUG oslo_concurrency.lockutils [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] Acquiring lock "625a3143-d138-4b52-aeb7-9e365c6f1128-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.154015] env[61985]: DEBUG oslo_concurrency.lockutils [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.154015] env[61985]: DEBUG oslo_concurrency.lockutils [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.154015] env[61985]: DEBUG nova.compute.manager [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] No waiting events found dispatching network-vif-plugged-262813ed-9baf-4240-b57d-fc4a648b5532 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 945.154015] env[61985]: WARNING nova.compute.manager [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Received unexpected event network-vif-plugged-262813ed-9baf-4240-b57d-fc4a648b5532 for instance with vm_state building and task_state spawning. [ 945.154015] env[61985]: DEBUG nova.compute.manager [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Received event network-changed-262813ed-9baf-4240-b57d-fc4a648b5532 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 945.154015] env[61985]: DEBUG nova.compute.manager [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Refreshing instance network info cache due to event network-changed-262813ed-9baf-4240-b57d-fc4a648b5532. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 945.154015] env[61985]: DEBUG oslo_concurrency.lockutils [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] Acquiring lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.167081] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "942e0f09-4d34-4aa2-8a60-4c12f899712c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.167352] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "942e0f09-4d34-4aa2-8a60-4c12f899712c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.220111] env[61985]: INFO nova.compute.resource_tracker [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating resource usage from migration 9da5e5c0-4f88-4a0e-98ee-ea219e2fc85d [ 945.288116] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.288542] env[61985]: DEBUG nova.compute.manager [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Instance network_info: |[{"id": "262813ed-9baf-4240-b57d-fc4a648b5532", "address": "fa:16:3e:d3:c1:6a", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262813ed-9b", "ovs_interfaceid": "262813ed-9baf-4240-b57d-fc4a648b5532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 945.288911] env[61985]: DEBUG oslo_concurrency.lockutils [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] Acquired lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.289186] env[61985]: DEBUG nova.network.neutron [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Refreshing network info cache for port 262813ed-9baf-4240-b57d-fc4a648b5532 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 945.291150] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:c1:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bff6c3a1-cc80-46ca-86c0-6dbb029edddb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '262813ed-9baf-4240-b57d-fc4a648b5532', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.300606] env[61985]: DEBUG oslo.service.loopingcall [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.305344] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 945.309805] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f8a72b7e-ffa8-490e-943a-9bddb804a553 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.334441] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936176, 'name': CloneVM_Task} progress is 95%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.343824] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.343824] env[61985]: value = "task-936178" [ 945.343824] env[61985]: _type = "Task" [ 945.343824] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.356108] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936178, 'name': CreateVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.605025] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ddafb74-0f88-4a0c-99a6-7e3e953e8ba2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.611698] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8dbd82a-801d-455f-9919-56614be2b5a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.644770] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1155b23-238d-48f6-b47a-39374bb1338f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.654616] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc0a9c7-fbf4-4e01-a4fb-094aee311002 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.671969] env[61985]: DEBUG nova.compute.provider_tree [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.673664] env[61985]: DEBUG nova.compute.manager [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 945.801564] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936176, 'name': CloneVM_Task, 'duration_secs': 1.901071} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.804280] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Created linked-clone VM from snapshot [ 945.805641] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c81467-8159-48b2-bebc-4024bae68e58 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.814498] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Uploading image c514448c-81e9-448f-b6fc-aac1a59a013c {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 945.829489] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 945.829776] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-103322f5-b30d-4835-bd4f-e4cbbd70a5b2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.836833] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 945.836833] env[61985]: value = "task-936179" [ 945.836833] env[61985]: _type = "Task" [ 945.836833] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.845971] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936179, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.854415] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936178, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.964287] env[61985]: DEBUG nova.network.neutron [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updating instance_info_cache with network_info: [{"id": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "address": "fa:16:3e:fc:45:f4", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97f641fb-39", "ovs_interfaceid": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.183856] env[61985]: DEBUG nova.scheduler.client.report [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.206666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.347588] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936179, 'name': Destroy_Task} progress is 33%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.355798] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936178, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.453283] env[61985]: DEBUG nova.network.neutron [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updated VIF entry in instance network info cache for port 262813ed-9baf-4240-b57d-fc4a648b5532. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 946.453726] env[61985]: DEBUG nova.network.neutron [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance_info_cache with network_info: [{"id": "262813ed-9baf-4240-b57d-fc4a648b5532", "address": "fa:16:3e:d3:c1:6a", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262813ed-9b", "ovs_interfaceid": "262813ed-9baf-4240-b57d-fc4a648b5532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.469190] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.694145] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.490s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.694421] env[61985]: INFO nova.compute.manager [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Migrating [ 946.707057] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.816s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.707285] env[61985]: DEBUG nova.objects.instance [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lazy-loading 'resources' on Instance uuid 3694e20c-ce37-4097-9991-8a06f38b2734 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.848633] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936179, 'name': Destroy_Task, 'duration_secs': 0.614652} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.853803] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Destroyed the VM [ 946.854216] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 946.855516] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3793826a-36ee-4a00-b9b1-c394240ef3f3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.863699] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936178, 'name': CreateVM_Task, 'duration_secs': 1.142384} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.865165] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 946.865674] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 946.865674] env[61985]: value = "task-936180" [ 946.865674] env[61985]: _type = "Task" [ 946.865674] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.866535] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.866713] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.867052] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 946.867359] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-987c6b70-94b2-4384-8a88-b9b2b9bc47fe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.878290] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 946.878290] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a93fed-072f-e907-e86a-2ceb08b93a5a" [ 946.878290] env[61985]: _type = "Task" [ 946.878290] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.883045] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936180, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.892019] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a93fed-072f-e907-e86a-2ceb08b93a5a, 'name': SearchDatastore_Task, 'duration_secs': 0.010238} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.892339] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.892586] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.892850] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.893021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.893215] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.893483] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-308e8d37-59c8-420e-83a8-9a480a2842ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.902537] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.902755] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 946.903495] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d5e2267-e43e-48c0-831d-b402722b3056 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.909409] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 946.909409] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d886c5-f246-c156-e1bd-84ef3aa74523" [ 946.909409] env[61985]: _type = "Task" [ 946.909409] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.917289] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d886c5-f246-c156-e1bd-84ef3aa74523, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.956230] env[61985]: DEBUG oslo_concurrency.lockutils [req-daa02199-2349-40a0-9b9a-c13e3519570e req-a277d08c-353d-41d8-9fc0-977622a2a99b service nova] Releasing lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.999303] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 946.999418] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16608ee1-71d8-4906-8627-704209856e1a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.010500] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 947.010500] env[61985]: value = "task-936181" [ 947.010500] env[61985]: _type = "Task" [ 947.010500] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.021701] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936181, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.213537] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.213771] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.213980] env[61985]: DEBUG nova.network.neutron [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 947.379620] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936180, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.396331] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.396331] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.422809] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d886c5-f246-c156-e1bd-84ef3aa74523, 'name': SearchDatastore_Task, 'duration_secs': 0.009096} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.425910] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c82df6ad-852d-439e-a4bd-026f592bcf44 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.434595] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 947.434595] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52effeac-6265-c423-f879-671de32efe91" [ 947.434595] env[61985]: _type = "Task" [ 947.434595] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.442927] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52effeac-6265-c423-f879-671de32efe91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.478238] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "2322ab72-9841-41fb-9d60-2812baabe108" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.478491] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.478683] env[61985]: DEBUG nova.compute.manager [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Going to confirm migration 2 {{(pid=61985) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4789}} [ 947.524104] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936181, 'name': PowerOffVM_Task, 'duration_secs': 0.278111} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.524395] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 947.525211] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f6d25d-f5a1-4c87-81d8-49704c7695b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.561499] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076205ee-594d-4812-9a89-899d556b3a3a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.596271] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 947.596604] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5906743c-0368-49f8-96b6-0d126caa6091 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.605129] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 947.605129] env[61985]: value = "task-936182" [ 947.605129] env[61985]: _type = "Task" [ 947.605129] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.613471] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936182, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.625767] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5f3f01-b120-4559-aa1b-d657f9e96977 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.633675] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1bf831-ade1-42b7-9fc2-0ac3cc1598f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.663667] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660a7504-9306-4c76-bcc1-8f47e3d6dc87 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.672343] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d948762-4972-4769-8f49-ba83c0a19766 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.687925] env[61985]: DEBUG nova.compute.provider_tree [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 947.882519] env[61985]: DEBUG oslo_vmware.api [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936180, 'name': RemoveSnapshot_Task, 'duration_secs': 0.7062} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.882871] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 947.898211] env[61985]: DEBUG nova.compute.manager [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 947.950432] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52effeac-6265-c423-f879-671de32efe91, 'name': SearchDatastore_Task, 'duration_secs': 0.010321} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.950830] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.950945] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 625a3143-d138-4b52-aeb7-9e365c6f1128/625a3143-d138-4b52-aeb7-9e365c6f1128.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 947.951215] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30cb659c-5757-467f-ab9e-6db7c6e6117d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.958805] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 947.958805] env[61985]: value = "task-936183" [ 947.958805] env[61985]: _type = "Task" [ 947.958805] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.970372] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936183, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.070669] env[61985]: DEBUG nova.network.neutron [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance_info_cache with network_info: [{"id": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "address": "fa:16:3e:b5:b4:b7", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.147", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d016adf-58", "ovs_interfaceid": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.077605] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.077849] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.078046] env[61985]: DEBUG nova.network.neutron [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 948.078247] env[61985]: DEBUG nova.objects.instance [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'info_cache' on Instance uuid 2322ab72-9841-41fb-9d60-2812baabe108 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.119548] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 948.119792] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.120065] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.120226] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.120413] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.120683] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15746936-4de6-4607-9b88-f32dbf1a6596 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.134432] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.134432] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 948.135672] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8bb6735-75b8-4ddb-8886-26c904e831f6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.142765] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 948.142765] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cf8cb9-6d03-2fe6-95a5-134c75df8c1b" [ 948.142765] env[61985]: _type = "Task" [ 948.142765] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.152999] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cf8cb9-6d03-2fe6-95a5-134c75df8c1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.212037] env[61985]: ERROR nova.scheduler.client.report [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] [req-8d55ea8b-f0b2-41fd-995e-8f7a197bb0ec] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8d55ea8b-f0b2-41fd-995e-8f7a197bb0ec"}]} [ 948.230461] env[61985]: DEBUG nova.scheduler.client.report [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 948.246353] env[61985]: DEBUG nova.scheduler.client.report [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 948.246624] env[61985]: DEBUG nova.compute.provider_tree [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 948.261129] env[61985]: DEBUG nova.scheduler.client.report [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 948.283291] env[61985]: DEBUG nova.scheduler.client.report [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 948.389704] env[61985]: WARNING nova.compute.manager [None req-6baa70a7-fb8c-43ae-b757-109133701fcb tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Image not found during snapshot: nova.exception.ImageNotFound: Image c514448c-81e9-448f-b6fc-aac1a59a013c could not be found. [ 948.425078] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.470507] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936183, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498412} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.470791] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 625a3143-d138-4b52-aeb7-9e365c6f1128/625a3143-d138-4b52-aeb7-9e365c6f1128.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 948.471020] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.471284] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69d8efdf-ead6-4aed-9051-9dc41a9a04aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.477760] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 948.477760] env[61985]: value = "task-936184" [ 948.477760] env[61985]: _type = "Task" [ 948.477760] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.486228] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936184, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.574131] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.631157] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c4689e-7347-4195-a8e0-d51fd04704c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.641289] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275ca70a-001a-49b5-acd5-9370ac4fc28b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.658788] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cf8cb9-6d03-2fe6-95a5-134c75df8c1b, 'name': SearchDatastore_Task, 'duration_secs': 0.019735} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.694541] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5c57e38-675d-4d90-8fca-5a5f0e35bd6e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.697197] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed91e58-8548-40ba-847e-445ce5ecab47 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.706263] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63ebd54-ab2d-473f-acc0-242869a1f5dc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.710198] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 948.710198] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521f7f92-9d0f-e9dc-b2c3-ece7076325c2" [ 948.710198] env[61985]: _type = "Task" [ 948.710198] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.720942] env[61985]: DEBUG nova.compute.provider_tree [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 948.727418] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521f7f92-9d0f-e9dc-b2c3-ece7076325c2, 'name': SearchDatastore_Task, 'duration_secs': 0.009851} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.727699] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.727967] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 6dd8218b-c14d-40c9-87df-097fab06c669/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk. {{(pid=61985) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 948.728239] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-297c3b17-7b37-4923-9b40-391d4e68db60 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.735586] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 948.735586] env[61985]: value = "task-936185" [ 948.735586] env[61985]: _type = "Task" [ 948.735586] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.743473] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936185, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.987970] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936184, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068511} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.988310] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.989120] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d27ea9-0b1f-4913-917f-2888da473284 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.014377] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 625a3143-d138-4b52-aeb7-9e365c6f1128/625a3143-d138-4b52-aeb7-9e365c6f1128.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.014721] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f78515e-ea3b-49a9-ba5e-4be7f295b6d5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.036609] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 949.036609] env[61985]: value = "task-936186" [ 949.036609] env[61985]: _type = "Task" [ 949.036609] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.046655] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936186, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.246979] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936185, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493613} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.247458] env[61985]: INFO nova.virt.vmwareapi.ds_util [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 6dd8218b-c14d-40c9-87df-097fab06c669/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk. [ 949.248159] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643f52f7-9f22-4405-8a96-bbd1fa511916 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.277495] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 6dd8218b-c14d-40c9-87df-097fab06c669/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.278681] env[61985]: DEBUG nova.scheduler.client.report [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 102 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 949.278927] env[61985]: DEBUG nova.compute.provider_tree [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 102 to 103 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 949.279181] env[61985]: DEBUG nova.compute.provider_tree [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 949.284458] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-965501a5-16c3-4f55-9b23-6a410f9e713b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.305711] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 949.305711] env[61985]: value = "task-936187" [ 949.305711] env[61985]: _type = "Task" [ 949.305711] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.316597] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936187, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.339014] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "488d0b95-97a9-4193-af62-cc92caf99625" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.339385] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "488d0b95-97a9-4193-af62-cc92caf99625" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.339636] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "488d0b95-97a9-4193-af62-cc92caf99625-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.339834] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "488d0b95-97a9-4193-af62-cc92caf99625-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.340023] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "488d0b95-97a9-4193-af62-cc92caf99625-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.342380] env[61985]: INFO nova.compute.manager [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Terminating instance [ 949.345806] env[61985]: DEBUG nova.compute.manager [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 949.346066] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.346935] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472a87a2-5029-4d3d-8961-1b8b41152f95 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.355531] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.355742] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fad1c1e1-5962-4617-b804-e299d11bb899 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.363139] env[61985]: DEBUG oslo_vmware.api [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 949.363139] env[61985]: value = "task-936188" [ 949.363139] env[61985]: _type = "Task" [ 949.363139] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.372807] env[61985]: DEBUG oslo_vmware.api [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.533355] env[61985]: DEBUG nova.network.neutron [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance_info_cache with network_info: [{"id": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "address": "fa:16:3e:12:45:ef", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aa18b7b-89", "ovs_interfaceid": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.546730] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936186, 'name': ReconfigVM_Task, 'duration_secs': 0.351039} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.547467] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 625a3143-d138-4b52-aeb7-9e365c6f1128/625a3143-d138-4b52-aeb7-9e365c6f1128.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.548946] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c929b2c5-44bf-4371-9e72-637897a1da2c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.557233] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 949.557233] env[61985]: value = "task-936189" [ 949.557233] env[61985]: _type = "Task" [ 949.557233] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.568470] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936189, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.800210] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.093s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.802579] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.019s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.804052] env[61985]: INFO nova.compute.claims [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.818108] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936187, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.819129] env[61985]: INFO nova.scheduler.client.report [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Deleted allocations for instance 3694e20c-ce37-4097-9991-8a06f38b2734 [ 949.873692] env[61985]: DEBUG oslo_vmware.api [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936188, 'name': PowerOffVM_Task, 'duration_secs': 0.194357} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.874032] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 949.874228] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 949.874693] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80a60aed-e2f3-43d6-9cbe-10fa17102312 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.955291] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 949.955568] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 949.955781] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Deleting the datastore file [datastore2] 488d0b95-97a9-4193-af62-cc92caf99625 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.956058] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa6e3f09-8b62-4616-828e-cf1722657fc6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.962813] env[61985]: DEBUG oslo_vmware.api [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 949.962813] env[61985]: value = "task-936191" [ 949.962813] env[61985]: _type = "Task" [ 949.962813] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.971010] env[61985]: DEBUG oslo_vmware.api [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936191, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.038548] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.039017] env[61985]: DEBUG nova.objects.instance [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'migration_context' on Instance uuid 2322ab72-9841-41fb-9d60-2812baabe108 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.068199] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936189, 'name': Rename_Task, 'duration_secs': 0.14742} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.068503] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 950.068762] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6afa4cab-e022-44a8-8ce8-a45e701410cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.077427] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 950.077427] env[61985]: value = "task-936192" [ 950.077427] env[61985]: _type = "Task" [ 950.077427] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.092347] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936192, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.093664] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ada7f9-311a-4021-bdfd-b65c4651532b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.113236] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance 'a77ca23f-b2c0-4822-8e48-3e47e0dadb27' progress to 0 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 950.318106] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936187, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.327512] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d9d1da41-c085-47a1-b398-5516bf8a1787 tempest-FloatingIPsAssociationTestJSON-1758581101 tempest-FloatingIPsAssociationTestJSON-1758581101-project-member] Lock "3694e20c-ce37-4097-9991-8a06f38b2734" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.591s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.474069] env[61985]: DEBUG oslo_vmware.api [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936191, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161542} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.474314] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.474509] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.474692] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.474876] env[61985]: INFO nova.compute.manager [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Took 1.13 seconds to destroy the instance on the hypervisor. [ 950.475144] env[61985]: DEBUG oslo.service.loopingcall [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.475343] env[61985]: DEBUG nova.compute.manager [-] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.475441] env[61985]: DEBUG nova.network.neutron [-] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.542463] env[61985]: DEBUG nova.objects.base [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Object Instance<2322ab72-9841-41fb-9d60-2812baabe108> lazy-loaded attributes: info_cache,migration_context {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 950.543517] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a0f76a-5457-4681-b898-2603606e6e55 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.566438] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddbb889e-5511-4c64-ae44-e8c6ed221768 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.572778] env[61985]: DEBUG oslo_vmware.api [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 950.572778] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d5f161-afa7-3479-6097-18c874d4a35a" [ 950.572778] env[61985]: _type = "Task" [ 950.572778] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.581738] env[61985]: DEBUG oslo_vmware.api [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d5f161-afa7-3479-6097-18c874d4a35a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.590019] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936192, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.620980] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 950.620980] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b684ae90-92dc-46b0-907c-98aaad010db8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.627919] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 950.627919] env[61985]: value = "task-936193" [ 950.627919] env[61985]: _type = "Task" [ 950.627919] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.637454] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936193, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.817299] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936187, 'name': ReconfigVM_Task, 'duration_secs': 1.416076} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.817722] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 6dd8218b-c14d-40c9-87df-097fab06c669/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.818636] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066df991-00fc-48c3-98e6-33dce04d4bdc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.853700] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af5a7874-bd08-401b-9425-fb72d6f118a0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.871796] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 950.871796] env[61985]: value = "task-936194" [ 950.871796] env[61985]: _type = "Task" [ 950.871796] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.883771] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936194, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.082941] env[61985]: DEBUG oslo_vmware.api [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d5f161-afa7-3479-6097-18c874d4a35a, 'name': SearchDatastore_Task, 'duration_secs': 0.00768} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.088517] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.096869] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936192, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.136685] env[61985]: DEBUG nova.compute.manager [req-15abace0-287d-4ecb-b247-295273802341 req-cd1473af-ff8d-4469-9051-b92718124dfa service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Received event network-vif-deleted-1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 951.137174] env[61985]: INFO nova.compute.manager [req-15abace0-287d-4ecb-b247-295273802341 req-cd1473af-ff8d-4469-9051-b92718124dfa service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Neutron deleted interface 1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4; detaching it from the instance and deleting it from the info cache [ 951.137350] env[61985]: DEBUG nova.network.neutron [req-15abace0-287d-4ecb-b247-295273802341 req-cd1473af-ff8d-4469-9051-b92718124dfa service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.143934] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936193, 'name': PowerOffVM_Task, 'duration_secs': 0.184242} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.144620] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 951.144818] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance 'a77ca23f-b2c0-4822-8e48-3e47e0dadb27' progress to 17 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 951.242030] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a70b5e-c161-43f6-a9f2-637ccd7f3860 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.250399] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ebb0af-339d-4c8b-865f-a63f905ed407 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.289735] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb07729-37e7-488f-aa93-035879035580 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.298885] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6435f2b1-9ae6-4d8d-8c96-bc9eb2870b13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.314360] env[61985]: DEBUG nova.compute.provider_tree [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.390311] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936194, 'name': ReconfigVM_Task, 'duration_secs': 0.20238} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.390311] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 951.390311] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc68ab23-57cc-47d9-ab39-0aba767f564f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.396015] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 951.396015] env[61985]: value = "task-936195" [ 951.396015] env[61985]: _type = "Task" [ 951.396015] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.405746] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936195, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.592770] env[61985]: DEBUG nova.network.neutron [-] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.594204] env[61985]: DEBUG oslo_vmware.api [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936192, 'name': PowerOnVM_Task, 'duration_secs': 1.128983} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.594568] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 951.594837] env[61985]: INFO nova.compute.manager [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Took 9.45 seconds to spawn the instance on the hypervisor. [ 951.595106] env[61985]: DEBUG nova.compute.manager [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.596015] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a83093-862b-4246-97cb-3adde98cd4b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.645170] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee943ec6-1696-4eb7-a5e8-2fac8cf559fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.650918] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.651222] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.651395] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.651624] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.651839] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.652138] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.652402] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.652581] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.653078] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.653364] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.653602] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.660536] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b9fb18a-8697-4939-ba37-48be38497013 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.677591] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b00648d-7665-4675-8707-bae84dc36528 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.699424] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 951.699424] env[61985]: value = "task-936196" [ 951.699424] env[61985]: _type = "Task" [ 951.699424] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.715737] env[61985]: DEBUG nova.compute.manager [req-15abace0-287d-4ecb-b247-295273802341 req-cd1473af-ff8d-4469-9051-b92718124dfa service nova] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Detach interface failed, port_id=1bfe0ade-ba0b-401a-a28e-088cf4e8a1b4, reason: Instance 488d0b95-97a9-4193-af62-cc92caf99625 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 951.724432] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936196, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.817454] env[61985]: DEBUG nova.scheduler.client.report [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.911868] env[61985]: DEBUG oslo_vmware.api [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936195, 'name': PowerOnVM_Task, 'duration_secs': 0.498819} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.913295] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 951.918412] env[61985]: DEBUG nova.compute.manager [None req-0e5726da-1e56-4218-9277-01d2b04ae5e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.919690] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ba2145-98b9-4da7-96bd-551379d98d5d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.100642] env[61985]: INFO nova.compute.manager [-] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Took 1.62 seconds to deallocate network for instance. [ 952.114185] env[61985]: INFO nova.compute.manager [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Took 41.47 seconds to build instance. [ 952.211113] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936196, 'name': ReconfigVM_Task, 'duration_secs': 0.415417} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.211508] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance 'a77ca23f-b2c0-4822-8e48-3e47e0dadb27' progress to 33 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 952.326053] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.326053] env[61985]: DEBUG nova.compute.manager [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 952.327282] env[61985]: DEBUG oslo_concurrency.lockutils [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.858s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.327699] env[61985]: DEBUG nova.objects.instance [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lazy-loading 'resources' on Instance uuid aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.612570] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.615145] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa558613-8d97-4768-a263-e89b7940eaac tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.984s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.721767] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:29:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='d8c8cc63-01d7-4c63-90a9-d468b4fc7cea',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1410428606',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 952.722077] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 952.722263] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.722461] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 952.722626] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.722785] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 952.723007] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 952.723348] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 952.724163] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 952.724163] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 952.724163] env[61985]: DEBUG nova.virt.hardware [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.731682] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Reconfiguring VM instance instance-00000047 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 952.732080] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff3a5f1a-51d1-4e9a-bf88-480c874ef7cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.756968] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 952.756968] env[61985]: value = "task-936197" [ 952.756968] env[61985]: _type = "Task" [ 952.756968] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.768897] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936197, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.832899] env[61985]: DEBUG nova.compute.utils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 952.834469] env[61985]: DEBUG nova.compute.manager [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 952.834649] env[61985]: DEBUG nova.network.neutron [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 952.896625] env[61985]: DEBUG nova.policy [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72dbd4382a5b4656a2ca6768e1b7edc6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be4f1b894ce04f37a0d106e93f4101fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 953.108042] env[61985]: INFO nova.compute.manager [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Unrescuing [ 953.108042] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.108042] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquired lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.108042] env[61985]: DEBUG nova.network.neutron [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 953.258719] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da31926a-4c04-4076-851c-b298ff1f3fc0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.275199] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf2abdb-2492-4401-a4a0-e0d08374460a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.281181] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936197, 'name': ReconfigVM_Task, 'duration_secs': 0.183326} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.281474] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Reconfigured VM instance instance-00000047 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 953.282652] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e53bf4e-62c6-4670-bec7-c3438e05f469 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.310910] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0115f45d-f8ca-4e0e-9772-6be5f7619ac3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.333594] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] a77ca23f-b2c0-4822-8e48-3e47e0dadb27/a77ca23f-b2c0-4822-8e48-3e47e0dadb27.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 953.334529] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7645cf2d-d9f1-49de-8af4-f8c48776c152 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.350018] env[61985]: DEBUG nova.compute.manager [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 953.354049] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d3dea3-bd5f-4376-8581-a4c73f55b7a0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.359752] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 953.359752] env[61985]: value = "task-936198" [ 953.359752] env[61985]: _type = "Task" [ 953.359752] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.377803] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936198, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.378327] env[61985]: DEBUG nova.compute.provider_tree [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.383822] env[61985]: DEBUG nova.network.neutron [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Successfully created port: 2ace42f2-9857-4a83-bcd2-75e3f0a3e788 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 953.872969] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936198, 'name': ReconfigVM_Task, 'duration_secs': 0.2513} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.873483] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Reconfigured VM instance instance-00000047 to attach disk [datastore1] a77ca23f-b2c0-4822-8e48-3e47e0dadb27/a77ca23f-b2c0-4822-8e48-3e47e0dadb27.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.873781] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance 'a77ca23f-b2c0-4822-8e48-3e47e0dadb27' progress to 50 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 953.882268] env[61985]: DEBUG nova.scheduler.client.report [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.181621] env[61985]: DEBUG nova.network.neutron [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updating instance_info_cache with network_info: [{"id": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "address": "fa:16:3e:fc:45:f4", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97f641fb-39", "ovs_interfaceid": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.307743] env[61985]: DEBUG nova.compute.manager [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Stashing vm_state: active {{(pid=61985) _prep_resize /opt/stack/nova/nova/compute/manager.py:5630}} [ 954.368824] env[61985]: DEBUG nova.compute.manager [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 954.380560] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16eed8f5-7758-442b-b4bb-9af8c64aa19c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.402265] env[61985]: DEBUG oslo_concurrency.lockutils [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.075s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.407512] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.903s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.409118] env[61985]: INFO nova.compute.claims [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.413021] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1429da2-09d9-4b3f-99ce-ec3a9beff926 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.436701] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance 'a77ca23f-b2c0-4822-8e48-3e47e0dadb27' progress to 67 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 954.441622] env[61985]: INFO nova.scheduler.client.report [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted allocations for instance aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3 [ 954.447524] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 954.447993] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 954.447993] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.448311] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 954.448480] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.448761] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 954.449131] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 954.449131] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 954.449318] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 954.449489] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 954.449995] env[61985]: DEBUG nova.virt.hardware [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.451033] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cba3441-b517-476e-979e-bbc3cff02915 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.459721] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8117aacf-f382-462d-ba34-808503ce0fd6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.684943] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Releasing lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.685947] env[61985]: DEBUG nova.objects.instance [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lazy-loading 'flavor' on Instance uuid 6dd8218b-c14d-40c9-87df-097fab06c669 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.834864] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.957937] env[61985]: DEBUG oslo_concurrency.lockutils [None req-71708779-06fd-431a-a267-2ba6ef638cd6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.630s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.980178] env[61985]: DEBUG nova.network.neutron [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Port 1d016adf-58f0-4fc2-a9d5-4e7232356db4 binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 955.193974] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a6c64b-bfc1-490b-8afb-04011c6368eb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.221146] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 955.221497] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42587d2c-e33a-4fd1-bf13-9dd85247c8cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.228913] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 955.228913] env[61985]: value = "task-936199" [ 955.228913] env[61985]: _type = "Task" [ 955.228913] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.237729] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.379996] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "49707589-4969-4e08-882b-2a2c94bc0d85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.380282] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "49707589-4969-4e08-882b-2a2c94bc0d85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.580041] env[61985]: DEBUG nova.network.neutron [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Successfully updated port: 2ace42f2-9857-4a83-bcd2-75e3f0a3e788 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.621731] env[61985]: DEBUG nova.compute.manager [req-d580a458-a58d-474b-94d6-05e23742a05b req-321d6dd1-1db3-4080-93d4-ac6a4115d09c service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Received event network-vif-plugged-2ace42f2-9857-4a83-bcd2-75e3f0a3e788 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 955.622773] env[61985]: DEBUG oslo_concurrency.lockutils [req-d580a458-a58d-474b-94d6-05e23742a05b req-321d6dd1-1db3-4080-93d4-ac6a4115d09c service nova] Acquiring lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.622773] env[61985]: DEBUG oslo_concurrency.lockutils [req-d580a458-a58d-474b-94d6-05e23742a05b req-321d6dd1-1db3-4080-93d4-ac6a4115d09c service nova] Lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.622773] env[61985]: DEBUG oslo_concurrency.lockutils [req-d580a458-a58d-474b-94d6-05e23742a05b req-321d6dd1-1db3-4080-93d4-ac6a4115d09c service nova] Lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.622773] env[61985]: DEBUG nova.compute.manager [req-d580a458-a58d-474b-94d6-05e23742a05b req-321d6dd1-1db3-4080-93d4-ac6a4115d09c service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] No waiting events found dispatching network-vif-plugged-2ace42f2-9857-4a83-bcd2-75e3f0a3e788 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.622773] env[61985]: WARNING nova.compute.manager [req-d580a458-a58d-474b-94d6-05e23742a05b req-321d6dd1-1db3-4080-93d4-ac6a4115d09c service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Received unexpected event network-vif-plugged-2ace42f2-9857-4a83-bcd2-75e3f0a3e788 for instance with vm_state building and task_state spawning. [ 955.741823] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936199, 'name': PowerOffVM_Task, 'duration_secs': 0.435826} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.742555] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 955.752440] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Reconfiguring VM instance instance-00000043 to detach disk 2002 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 955.756988] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edcd8a29-3eca-49a6-b55a-ecb3262557a8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.785847] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 955.785847] env[61985]: value = "task-936200" [ 955.785847] env[61985]: _type = "Task" [ 955.785847] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.794683] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936200, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.883283] env[61985]: DEBUG nova.compute.manager [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 955.913652] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358a6743-66bb-4f69-a92f-1dc1b8f5707a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.921488] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad31a2b-ec0f-458c-bda3-3f0256c21b55 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.953526] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d5874a-0a95-498b-9ab7-104020d0db44 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.961126] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa614d6-e407-424e-ac61-7701abab5fd8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.975435] env[61985]: DEBUG nova.compute.provider_tree [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.006062] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.006370] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.006568] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.085861] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "refresh_cache-57591f6f-ed7d-46f3-a9c1-217016c42b1c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.085861] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "refresh_cache-57591f6f-ed7d-46f3-a9c1-217016c42b1c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.085861] env[61985]: DEBUG nova.network.neutron [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 956.297118] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936200, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.403410] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.479033] env[61985]: DEBUG nova.scheduler.client.report [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.636567] env[61985]: DEBUG nova.network.neutron [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 956.797180] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936200, 'name': ReconfigVM_Task, 'duration_secs': 0.703742} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.801022] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Reconfigured VM instance instance-00000043 to detach disk 2002 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 956.801022] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 956.801022] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-363391e1-0423-4bc2-9e90-3d3c87fb14cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.804617] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 956.804617] env[61985]: value = "task-936201" [ 956.804617] env[61985]: _type = "Task" [ 956.804617] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.819973] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936201, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.887249] env[61985]: DEBUG nova.network.neutron [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Updating instance_info_cache with network_info: [{"id": "2ace42f2-9857-4a83-bcd2-75e3f0a3e788", "address": "fa:16:3e:a8:2c:bb", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ace42f2-98", "ovs_interfaceid": "2ace42f2-9857-4a83-bcd2-75e3f0a3e788", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.984200] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.576s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.984936] env[61985]: DEBUG nova.compute.manager [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 956.992021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.751s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.992021] env[61985]: INFO nova.compute.claims [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.062724] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.062927] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.063250] env[61985]: DEBUG nova.network.neutron [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 957.114130] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.114374] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.314480] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936201, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.393069] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "refresh_cache-57591f6f-ed7d-46f3-a9c1-217016c42b1c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.393069] env[61985]: DEBUG nova.compute.manager [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Instance network_info: |[{"id": "2ace42f2-9857-4a83-bcd2-75e3f0a3e788", "address": "fa:16:3e:a8:2c:bb", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ace42f2-98", "ovs_interfaceid": "2ace42f2-9857-4a83-bcd2-75e3f0a3e788", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 957.393069] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:2c:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cb0e556a-0f69-4a5c-af62-ffc46edb8e63', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ace42f2-9857-4a83-bcd2-75e3f0a3e788', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.402613] env[61985]: DEBUG oslo.service.loopingcall [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.402998] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 957.404086] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4fe0265f-c2e8-4095-9a8e-0fd35692f70e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.424952] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.424952] env[61985]: value = "task-936202" [ 957.424952] env[61985]: _type = "Task" [ 957.424952] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.432398] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936202, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.499301] env[61985]: DEBUG nova.compute.utils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 957.501042] env[61985]: DEBUG nova.compute.manager [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 957.501369] env[61985]: DEBUG nova.network.neutron [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 957.540638] env[61985]: DEBUG nova.policy [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f903c880e12c4dbbb87c2a1d3edaaf53', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f99d7356c6d45eb8442ad923ab548b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.616455] env[61985]: DEBUG nova.compute.manager [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 957.655259] env[61985]: DEBUG nova.compute.manager [req-c18e32be-1040-4ea4-acfb-05fa6923f24c req-eef209f0-aaf7-4da2-9a3c-1560233681f4 service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Received event network-changed-2ace42f2-9857-4a83-bcd2-75e3f0a3e788 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 957.655471] env[61985]: DEBUG nova.compute.manager [req-c18e32be-1040-4ea4-acfb-05fa6923f24c req-eef209f0-aaf7-4da2-9a3c-1560233681f4 service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Refreshing instance network info cache due to event network-changed-2ace42f2-9857-4a83-bcd2-75e3f0a3e788. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 957.655909] env[61985]: DEBUG oslo_concurrency.lockutils [req-c18e32be-1040-4ea4-acfb-05fa6923f24c req-eef209f0-aaf7-4da2-9a3c-1560233681f4 service nova] Acquiring lock "refresh_cache-57591f6f-ed7d-46f3-a9c1-217016c42b1c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.656854] env[61985]: DEBUG oslo_concurrency.lockutils [req-c18e32be-1040-4ea4-acfb-05fa6923f24c req-eef209f0-aaf7-4da2-9a3c-1560233681f4 service nova] Acquired lock "refresh_cache-57591f6f-ed7d-46f3-a9c1-217016c42b1c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.656854] env[61985]: DEBUG nova.network.neutron [req-c18e32be-1040-4ea4-acfb-05fa6923f24c req-eef209f0-aaf7-4da2-9a3c-1560233681f4 service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Refreshing network info cache for port 2ace42f2-9857-4a83-bcd2-75e3f0a3e788 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 957.814249] env[61985]: DEBUG oslo_vmware.api [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936201, 'name': PowerOnVM_Task, 'duration_secs': 1.004444} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.814537] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 957.814774] env[61985]: DEBUG nova.compute.manager [None req-a92ff32b-af8a-4a3c-bdd0-70db6f90107a tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.815567] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f898c0-a2c7-4742-8247-cb417113fe34 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.934537] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936202, 'name': CreateVM_Task, 'duration_secs': 0.310149} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.934712] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 957.935687] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.935687] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.935902] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.936172] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3929a014-98ac-473b-8592-edcb7babc7ea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.940624] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 957.940624] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52161413-24ba-65c0-643b-a4ae265f75de" [ 957.940624] env[61985]: _type = "Task" [ 957.940624] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.948500] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52161413-24ba-65c0-643b-a4ae265f75de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.004556] env[61985]: DEBUG nova.compute.manager [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 958.059779] env[61985]: DEBUG nova.network.neutron [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance_info_cache with network_info: [{"id": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "address": "fa:16:3e:b5:b4:b7", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.147", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d016adf-58", "ovs_interfaceid": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.136657] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.240869] env[61985]: DEBUG nova.network.neutron [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Successfully created port: 92893cd3-94c2-4842-a021-372bd8d06fda {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 958.384504] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33669d6-fdd6-4df8-a624-2c2738a3b827 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.392224] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d083388-78e7-4ea9-a341-5f60d6e9ef75 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.429670] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd43c8f-5859-4bfe-8194-e84e130b0221 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.438119] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119a69f8-f609-41a5-a8dd-dc6c784faabe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.455218] env[61985]: DEBUG nova.compute.provider_tree [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.459413] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52161413-24ba-65c0-643b-a4ae265f75de, 'name': SearchDatastore_Task, 'duration_secs': 0.009635} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.460009] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.460538] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.460538] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.460720] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.460810] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.461072] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-68330baf-2faf-4b51-a8f2-de3f245a46c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.470091] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.470332] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 958.471092] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-734e71e5-7098-49d4-a1a1-c69a84eb1ba7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.476098] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 958.476098] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f111b3-39bb-5548-b4dd-4149027a541c" [ 958.476098] env[61985]: _type = "Task" [ 958.476098] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.484179] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f111b3-39bb-5548-b4dd-4149027a541c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.562362] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.724761] env[61985]: DEBUG nova.network.neutron [req-c18e32be-1040-4ea4-acfb-05fa6923f24c req-eef209f0-aaf7-4da2-9a3c-1560233681f4 service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Updated VIF entry in instance network info cache for port 2ace42f2-9857-4a83-bcd2-75e3f0a3e788. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 958.725260] env[61985]: DEBUG nova.network.neutron [req-c18e32be-1040-4ea4-acfb-05fa6923f24c req-eef209f0-aaf7-4da2-9a3c-1560233681f4 service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Updating instance_info_cache with network_info: [{"id": "2ace42f2-9857-4a83-bcd2-75e3f0a3e788", "address": "fa:16:3e:a8:2c:bb", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ace42f2-98", "ovs_interfaceid": "2ace42f2-9857-4a83-bcd2-75e3f0a3e788", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.961732] env[61985]: DEBUG nova.scheduler.client.report [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.989534] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f111b3-39bb-5548-b4dd-4149027a541c, 'name': SearchDatastore_Task, 'duration_secs': 0.011385} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.990818] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-baf7604a-465e-4ca3-b626-fdfc065e673f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.998521] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 958.998521] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a6c96c-4e80-295e-b9d9-14a0240a8dbe" [ 958.998521] env[61985]: _type = "Task" [ 958.998521] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.011075] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a6c96c-4e80-295e-b9d9-14a0240a8dbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.013518] env[61985]: DEBUG nova.compute.manager [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 959.040593] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.040922] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.041170] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.041441] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.041664] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.041894] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.042211] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.042453] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.042725] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.042991] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.043256] env[61985]: DEBUG nova.virt.hardware [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.044387] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba10a6e6-8408-4a38-8971-4185ef75fcc9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.055645] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed78b8f7-92cc-41a0-b1b9-8eafc902073b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.094917] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9b698f-4d2d-4b82-a347-4174b48c2b38 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.120914] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298b0963-dfa1-45c3-a6bc-830565700c79 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.129786] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance 'a77ca23f-b2c0-4822-8e48-3e47e0dadb27' progress to 83 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 959.228079] env[61985]: DEBUG oslo_concurrency.lockutils [req-c18e32be-1040-4ea4-acfb-05fa6923f24c req-eef209f0-aaf7-4da2-9a3c-1560233681f4 service nova] Releasing lock "refresh_cache-57591f6f-ed7d-46f3-a9c1-217016c42b1c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.469388] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.469980] env[61985]: DEBUG nova.compute.manager [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 959.472883] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.456s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.473136] env[61985]: DEBUG nova.objects.instance [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lazy-loading 'resources' on Instance uuid 5c2a7072-5b14-4ec7-8060-4e21eea927e3 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.508787] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a6c96c-4e80-295e-b9d9-14a0240a8dbe, 'name': SearchDatastore_Task, 'duration_secs': 0.014115} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.509745] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.510021] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 57591f6f-ed7d-46f3-a9c1-217016c42b1c/57591f6f-ed7d-46f3-a9c1-217016c42b1c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 959.510282] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e6104ea-c659-409f-8775-d6b758654022 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.516684] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 959.516684] env[61985]: value = "task-936203" [ 959.516684] env[61985]: _type = "Task" [ 959.516684] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.524160] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936203, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.620365] env[61985]: DEBUG nova.compute.manager [req-15258c04-afae-4335-9fbe-7f3f518e55ab req-f73b8eae-3ee9-4b62-a99f-d1806ae62cd6 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Received event network-changed-97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 959.621042] env[61985]: DEBUG nova.compute.manager [req-15258c04-afae-4335-9fbe-7f3f518e55ab req-f73b8eae-3ee9-4b62-a99f-d1806ae62cd6 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Refreshing instance network info cache due to event network-changed-97f641fb-39ac-441c-bacc-6705d1ea6e98. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 959.621512] env[61985]: DEBUG oslo_concurrency.lockutils [req-15258c04-afae-4335-9fbe-7f3f518e55ab req-f73b8eae-3ee9-4b62-a99f-d1806ae62cd6 service nova] Acquiring lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.622575] env[61985]: DEBUG oslo_concurrency.lockutils [req-15258c04-afae-4335-9fbe-7f3f518e55ab req-f73b8eae-3ee9-4b62-a99f-d1806ae62cd6 service nova] Acquired lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.622575] env[61985]: DEBUG nova.network.neutron [req-15258c04-afae-4335-9fbe-7f3f518e55ab req-f73b8eae-3ee9-4b62-a99f-d1806ae62cd6 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Refreshing network info cache for port 97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 959.636675] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 959.637036] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f461c5e7-3755-4615-a062-b2c5bce0ebab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.644676] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 959.644676] env[61985]: value = "task-936204" [ 959.644676] env[61985]: _type = "Task" [ 959.644676] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.652895] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936204, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.980157] env[61985]: DEBUG nova.compute.utils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 959.981931] env[61985]: DEBUG nova.compute.manager [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 959.982127] env[61985]: DEBUG nova.network.neutron [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 960.038017] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936203, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.109663] env[61985]: DEBUG nova.policy [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4f5b5ce76e8b4fae9c905bc7a39b683e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4923009773c047519b883745c59e3516', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 960.154993] env[61985]: DEBUG oslo_vmware.api [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936204, 'name': PowerOnVM_Task, 'duration_secs': 0.421273} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.158304] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 960.158693] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4337d657-0aab-4665-b3c1-030091cbb0d2 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance 'a77ca23f-b2c0-4822-8e48-3e47e0dadb27' progress to 100 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 960.335158] env[61985]: DEBUG nova.compute.manager [req-d50fd87c-0771-4fd0-9d08-68b6c4049763 req-9f1dc3cb-e120-4ce5-9b19-ac280440da99 service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Received event network-vif-plugged-92893cd3-94c2-4842-a021-372bd8d06fda {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 960.335158] env[61985]: DEBUG oslo_concurrency.lockutils [req-d50fd87c-0771-4fd0-9d08-68b6c4049763 req-9f1dc3cb-e120-4ce5-9b19-ac280440da99 service nova] Acquiring lock "4da2d122-aa36-4cad-9521-498b53b9f2f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.335158] env[61985]: DEBUG oslo_concurrency.lockutils [req-d50fd87c-0771-4fd0-9d08-68b6c4049763 req-9f1dc3cb-e120-4ce5-9b19-ac280440da99 service nova] Lock "4da2d122-aa36-4cad-9521-498b53b9f2f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.335158] env[61985]: DEBUG oslo_concurrency.lockutils [req-d50fd87c-0771-4fd0-9d08-68b6c4049763 req-9f1dc3cb-e120-4ce5-9b19-ac280440da99 service nova] Lock "4da2d122-aa36-4cad-9521-498b53b9f2f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.335158] env[61985]: DEBUG nova.compute.manager [req-d50fd87c-0771-4fd0-9d08-68b6c4049763 req-9f1dc3cb-e120-4ce5-9b19-ac280440da99 service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] No waiting events found dispatching network-vif-plugged-92893cd3-94c2-4842-a021-372bd8d06fda {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.335567] env[61985]: WARNING nova.compute.manager [req-d50fd87c-0771-4fd0-9d08-68b6c4049763 req-9f1dc3cb-e120-4ce5-9b19-ac280440da99 service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Received unexpected event network-vif-plugged-92893cd3-94c2-4842-a021-372bd8d06fda for instance with vm_state building and task_state spawning. [ 960.366947] env[61985]: DEBUG nova.network.neutron [req-15258c04-afae-4335-9fbe-7f3f518e55ab req-f73b8eae-3ee9-4b62-a99f-d1806ae62cd6 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updated VIF entry in instance network info cache for port 97f641fb-39ac-441c-bacc-6705d1ea6e98. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 960.366947] env[61985]: DEBUG nova.network.neutron [req-15258c04-afae-4335-9fbe-7f3f518e55ab req-f73b8eae-3ee9-4b62-a99f-d1806ae62cd6 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updating instance_info_cache with network_info: [{"id": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "address": "fa:16:3e:fc:45:f4", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97f641fb-39", "ovs_interfaceid": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.425876] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e909554-3ef1-4cb5-9a9d-fd30d45a8936 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.433797] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3a2fd0-6f97-4e77-a3cd-0f0d81f5bce3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.464420] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe0e65c-e9d3-414e-9aa8-297ac31dc025 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.472498] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c7b03f-cd34-4fa3-b0ee-1311fe4d60fb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.486645] env[61985]: DEBUG nova.compute.manager [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 960.489754] env[61985]: DEBUG nova.compute.provider_tree [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.536753] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936203, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.622309} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.537138] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 57591f6f-ed7d-46f3-a9c1-217016c42b1c/57591f6f-ed7d-46f3-a9c1-217016c42b1c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 960.537354] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.537683] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cee99143-acea-4af0-826a-cb2d4954757b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.545860] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 960.545860] env[61985]: value = "task-936205" [ 960.545860] env[61985]: _type = "Task" [ 960.545860] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.555395] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936205, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.645226] env[61985]: DEBUG nova.network.neutron [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Successfully updated port: 92893cd3-94c2-4842-a021-372bd8d06fda {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 960.779948] env[61985]: DEBUG nova.network.neutron [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Successfully created port: f33869c6-9fa7-44ab-8de2-ae92191b1096 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 960.874581] env[61985]: DEBUG oslo_concurrency.lockutils [req-15258c04-afae-4335-9fbe-7f3f518e55ab req-f73b8eae-3ee9-4b62-a99f-d1806ae62cd6 service nova] Releasing lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.995640] env[61985]: DEBUG nova.scheduler.client.report [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.059013] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936205, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.164913} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.059320] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 961.060114] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21c0b02-9fd0-458f-9f0d-f2321c7d4f72 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.086405] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 57591f6f-ed7d-46f3-a9c1-217016c42b1c/57591f6f-ed7d-46f3-a9c1-217016c42b1c.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.087350] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d080d7a-1482-4531-87a3-a88deb1e632c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.113019] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 961.113019] env[61985]: value = "task-936206" [ 961.113019] env[61985]: _type = "Task" [ 961.113019] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.121544] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936206, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.147456] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquiring lock "refresh_cache-4da2d122-aa36-4cad-9521-498b53b9f2f6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.147636] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquired lock "refresh_cache-4da2d122-aa36-4cad-9521-498b53b9f2f6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.147779] env[61985]: DEBUG nova.network.neutron [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.501030] env[61985]: DEBUG nova.compute.manager [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 961.504123] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.031s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.506218] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.880s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.506701] env[61985]: DEBUG nova.objects.instance [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lazy-loading 'resources' on Instance uuid 518635bf-73ee-404b-ae6a-dc4ee23009d2 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.526612] env[61985]: INFO nova.scheduler.client.report [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted allocations for instance 5c2a7072-5b14-4ec7-8060-4e21eea927e3 [ 961.534359] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 961.534359] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 961.534359] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.534359] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 961.534359] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.534359] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 961.534678] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 961.534770] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 961.534933] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 961.535123] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 961.535310] env[61985]: DEBUG nova.virt.hardware [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.536444] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efa4145-555f-44b0-8407-c3861f12a95f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.545314] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d859b4ce-2cc5-4669-9737-496f105650c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.619601] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936206, 'name': ReconfigVM_Task, 'duration_secs': 0.281971} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.619947] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 57591f6f-ed7d-46f3-a9c1-217016c42b1c/57591f6f-ed7d-46f3-a9c1-217016c42b1c.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.620590] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6255d80f-e4bc-46bb-abdf-4b02141bcd8c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.629433] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 961.629433] env[61985]: value = "task-936207" [ 961.629433] env[61985]: _type = "Task" [ 961.629433] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.636428] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936207, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.667521] env[61985]: DEBUG nova.compute.manager [req-0d413c1c-b8a6-428b-88f9-702596f6a25e req-b35dd6ad-1e56-4363-bace-1ce443404c5d service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Received event network-changed-97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 961.667934] env[61985]: DEBUG nova.compute.manager [req-0d413c1c-b8a6-428b-88f9-702596f6a25e req-b35dd6ad-1e56-4363-bace-1ce443404c5d service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Refreshing instance network info cache due to event network-changed-97f641fb-39ac-441c-bacc-6705d1ea6e98. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 961.668279] env[61985]: DEBUG oslo_concurrency.lockutils [req-0d413c1c-b8a6-428b-88f9-702596f6a25e req-b35dd6ad-1e56-4363-bace-1ce443404c5d service nova] Acquiring lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.668466] env[61985]: DEBUG oslo_concurrency.lockutils [req-0d413c1c-b8a6-428b-88f9-702596f6a25e req-b35dd6ad-1e56-4363-bace-1ce443404c5d service nova] Acquired lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.668693] env[61985]: DEBUG nova.network.neutron [req-0d413c1c-b8a6-428b-88f9-702596f6a25e req-b35dd6ad-1e56-4363-bace-1ce443404c5d service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Refreshing network info cache for port 97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 961.684974] env[61985]: DEBUG nova.network.neutron [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 961.944755] env[61985]: DEBUG nova.network.neutron [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Updating instance_info_cache with network_info: [{"id": "92893cd3-94c2-4842-a021-372bd8d06fda", "address": "fa:16:3e:b5:86:a2", "network": {"id": "f565ab3c-eb41-4e4e-aa35-1aa5def2ae23", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1127477073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f99d7356c6d45eb8442ad923ab548b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92893cd3-94", "ovs_interfaceid": "92893cd3-94c2-4842-a021-372bd8d06fda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.045344] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7e410d69-9ba7-4b6e-93b0-0c0453bb8944 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5c2a7072-5b14-4ec7-8060-4e21eea927e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.604s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.138396] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936207, 'name': Rename_Task, 'duration_secs': 0.133317} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.141212] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 962.141703] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7bb3b12-599a-4b76-a0fc-89f293765da5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.148566] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 962.148566] env[61985]: value = "task-936208" [ 962.148566] env[61985]: _type = "Task" [ 962.148566] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.158845] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936208, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.400445] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f1942f-14d1-4a50-8fc0-5b650863bd0b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.411416] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ebae55e-d6eb-46ff-ac27-34a77c3df16a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.449720] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Releasing lock "refresh_cache-4da2d122-aa36-4cad-9521-498b53b9f2f6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.450172] env[61985]: DEBUG nova.compute.manager [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Instance network_info: |[{"id": "92893cd3-94c2-4842-a021-372bd8d06fda", "address": "fa:16:3e:b5:86:a2", "network": {"id": "f565ab3c-eb41-4e4e-aa35-1aa5def2ae23", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1127477073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f99d7356c6d45eb8442ad923ab548b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92893cd3-94", "ovs_interfaceid": "92893cd3-94c2-4842-a021-372bd8d06fda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 962.454372] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:86:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b7a73c01-1bb9-4612-a1a7-16d71b732e81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92893cd3-94c2-4842-a021-372bd8d06fda', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.462624] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Creating folder: Project (3f99d7356c6d45eb8442ad923ab548b7). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 962.463416] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1463886b-9604-44a2-a090-f83cad5b941f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.473951] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-970163d4-dc65-453f-86ec-5a3ae73550e0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.483640] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-786134a1-9cc0-4cd4-9795-176b286710a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.492099] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Created folder: Project (3f99d7356c6d45eb8442ad923ab548b7) in parent group-v211285. [ 962.492540] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Creating folder: Instances. Parent ref: group-v211427. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 962.493198] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22b44ebc-53fc-42dd-b9bd-d5f811764e34 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.504567] env[61985]: DEBUG nova.compute.provider_tree [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.514881] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Created folder: Instances in parent group-v211427. [ 962.515740] env[61985]: DEBUG oslo.service.loopingcall [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.515740] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 962.515740] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-93d24c9f-5674-4992-8421-fdbb2b066496 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.535780] env[61985]: DEBUG nova.compute.manager [req-5e38a396-3893-4f33-ae76-64584595eeb0 req-1690949e-717e-4bc3-a969-b314ebd515df service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Received event network-changed-92893cd3-94c2-4842-a021-372bd8d06fda {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 962.536142] env[61985]: DEBUG nova.compute.manager [req-5e38a396-3893-4f33-ae76-64584595eeb0 req-1690949e-717e-4bc3-a969-b314ebd515df service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Refreshing instance network info cache due to event network-changed-92893cd3-94c2-4842-a021-372bd8d06fda. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 962.536260] env[61985]: DEBUG oslo_concurrency.lockutils [req-5e38a396-3893-4f33-ae76-64584595eeb0 req-1690949e-717e-4bc3-a969-b314ebd515df service nova] Acquiring lock "refresh_cache-4da2d122-aa36-4cad-9521-498b53b9f2f6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.536428] env[61985]: DEBUG oslo_concurrency.lockutils [req-5e38a396-3893-4f33-ae76-64584595eeb0 req-1690949e-717e-4bc3-a969-b314ebd515df service nova] Acquired lock "refresh_cache-4da2d122-aa36-4cad-9521-498b53b9f2f6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.536575] env[61985]: DEBUG nova.network.neutron [req-5e38a396-3893-4f33-ae76-64584595eeb0 req-1690949e-717e-4bc3-a969-b314ebd515df service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Refreshing network info cache for port 92893cd3-94c2-4842-a021-372bd8d06fda {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.543903] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.543903] env[61985]: value = "task-936211" [ 962.543903] env[61985]: _type = "Task" [ 962.543903] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.555053] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936211, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.661565] env[61985]: DEBUG oslo_vmware.api [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936208, 'name': PowerOnVM_Task, 'duration_secs': 0.448812} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.661846] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 962.662067] env[61985]: INFO nova.compute.manager [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Took 8.29 seconds to spawn the instance on the hypervisor. [ 962.662280] env[61985]: DEBUG nova.compute.manager [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.663451] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fbb6a1-1c28-4248-ad4d-6e7b21a7e6a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.824807] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "b6eb50ff-e685-4e8a-92f5-c2661136c361" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.825043] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "b6eb50ff-e685-4e8a-92f5-c2661136c361" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.008503] env[61985]: DEBUG nova.scheduler.client.report [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.011801] env[61985]: DEBUG nova.network.neutron [req-0d413c1c-b8a6-428b-88f9-702596f6a25e req-b35dd6ad-1e56-4363-bace-1ce443404c5d service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updated VIF entry in instance network info cache for port 97f641fb-39ac-441c-bacc-6705d1ea6e98. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 963.015020] env[61985]: DEBUG nova.network.neutron [req-0d413c1c-b8a6-428b-88f9-702596f6a25e req-b35dd6ad-1e56-4363-bace-1ce443404c5d service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updating instance_info_cache with network_info: [{"id": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "address": "fa:16:3e:fc:45:f4", "network": {"id": "dcfe138c-4918-4953-aa6e-c58732686f23", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1680478779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "091a92048e3545b7a6b98af4e81d2f00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97f641fb-39", "ovs_interfaceid": "97f641fb-39ac-441c-bacc-6705d1ea6e98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.054700] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936211, 'name': CreateVM_Task, 'duration_secs': 0.457995} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.054894] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 963.055679] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.055883] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.056306] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.056621] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32c8c453-65d2-4c67-a034-05f8b2d8a7d9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.061599] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for the task: (returnval){ [ 963.061599] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520cc040-16a4-53a6-2f23-4e0e5ca85c0a" [ 963.061599] env[61985]: _type = "Task" [ 963.061599] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.073486] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520cc040-16a4-53a6-2f23-4e0e5ca85c0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.095529] env[61985]: DEBUG nova.compute.manager [req-31d76f12-3072-4190-8f67-81f678652e07 req-f00e1917-2035-4f38-8995-b69d5f4c21e8 service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Received event network-vif-plugged-f33869c6-9fa7-44ab-8de2-ae92191b1096 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 963.095762] env[61985]: DEBUG oslo_concurrency.lockutils [req-31d76f12-3072-4190-8f67-81f678652e07 req-f00e1917-2035-4f38-8995-b69d5f4c21e8 service nova] Acquiring lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.096073] env[61985]: DEBUG oslo_concurrency.lockutils [req-31d76f12-3072-4190-8f67-81f678652e07 req-f00e1917-2035-4f38-8995-b69d5f4c21e8 service nova] Lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.096289] env[61985]: DEBUG oslo_concurrency.lockutils [req-31d76f12-3072-4190-8f67-81f678652e07 req-f00e1917-2035-4f38-8995-b69d5f4c21e8 service nova] Lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.096447] env[61985]: DEBUG nova.compute.manager [req-31d76f12-3072-4190-8f67-81f678652e07 req-f00e1917-2035-4f38-8995-b69d5f4c21e8 service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] No waiting events found dispatching network-vif-plugged-f33869c6-9fa7-44ab-8de2-ae92191b1096 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.096647] env[61985]: WARNING nova.compute.manager [req-31d76f12-3072-4190-8f67-81f678652e07 req-f00e1917-2035-4f38-8995-b69d5f4c21e8 service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Received unexpected event network-vif-plugged-f33869c6-9fa7-44ab-8de2-ae92191b1096 for instance with vm_state building and task_state spawning. [ 963.109114] env[61985]: DEBUG nova.network.neutron [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Port 1d016adf-58f0-4fc2-a9d5-4e7232356db4 binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 963.109410] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.109572] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.109742] env[61985]: DEBUG nova.network.neutron [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 963.184223] env[61985]: INFO nova.compute.manager [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Took 31.49 seconds to build instance. [ 963.327379] env[61985]: DEBUG nova.compute.manager [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 963.412272] env[61985]: DEBUG nova.network.neutron [req-5e38a396-3893-4f33-ae76-64584595eeb0 req-1690949e-717e-4bc3-a969-b314ebd515df service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Updated VIF entry in instance network info cache for port 92893cd3-94c2-4842-a021-372bd8d06fda. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 963.412272] env[61985]: DEBUG nova.network.neutron [req-5e38a396-3893-4f33-ae76-64584595eeb0 req-1690949e-717e-4bc3-a969-b314ebd515df service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Updating instance_info_cache with network_info: [{"id": "92893cd3-94c2-4842-a021-372bd8d06fda", "address": "fa:16:3e:b5:86:a2", "network": {"id": "f565ab3c-eb41-4e4e-aa35-1aa5def2ae23", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1127477073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f99d7356c6d45eb8442ad923ab548b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b7a73c01-1bb9-4612-a1a7-16d71b732e81", "external-id": "nsx-vlan-transportzone-711", "segmentation_id": 711, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92893cd3-94", "ovs_interfaceid": "92893cd3-94c2-4842-a021-372bd8d06fda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.468251] env[61985]: DEBUG nova.network.neutron [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Successfully updated port: f33869c6-9fa7-44ab-8de2-ae92191b1096 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.505750] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 963.506119] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 963.514477] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.008s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.516717] env[61985]: DEBUG oslo_concurrency.lockutils [req-0d413c1c-b8a6-428b-88f9-702596f6a25e req-b35dd6ad-1e56-4363-bace-1ce443404c5d service nova] Releasing lock "refresh_cache-6dd8218b-c14d-40c9-87df-097fab06c669" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.517552] env[61985]: DEBUG oslo_concurrency.lockutils [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.948s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.517813] env[61985]: DEBUG nova.objects.instance [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lazy-loading 'resources' on Instance uuid cd8e2eaf-3c13-452b-b2e6-8107f0219378 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.542999] env[61985]: INFO nova.scheduler.client.report [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Deleted allocations for instance 518635bf-73ee-404b-ae6a-dc4ee23009d2 [ 963.573177] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520cc040-16a4-53a6-2f23-4e0e5ca85c0a, 'name': SearchDatastore_Task, 'duration_secs': 0.015922} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.573355] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.573602] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.573881] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.574055] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.574249] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.574597] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06b299fd-9987-4c6f-a3c5-9143a9056d59 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.583360] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.583554] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 963.584293] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-304aae6c-5def-4214-ac1c-80efd4507c3a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.590159] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for the task: (returnval){ [ 963.590159] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529b1144-8dbb-bacf-8ba3-cb386f97416d" [ 963.590159] env[61985]: _type = "Task" [ 963.590159] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.598356] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529b1144-8dbb-bacf-8ba3-cb386f97416d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.684294] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1e5a373a-868b-4638-9934-e3f8f8176e45 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.005s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.805191] env[61985]: DEBUG nova.network.neutron [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance_info_cache with network_info: [{"id": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "address": "fa:16:3e:b5:b4:b7", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.147", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d016adf-58", "ovs_interfaceid": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.849703] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.913965] env[61985]: DEBUG oslo_concurrency.lockutils [req-5e38a396-3893-4f33-ae76-64584595eeb0 req-1690949e-717e-4bc3-a969-b314ebd515df service nova] Releasing lock "refresh_cache-4da2d122-aa36-4cad-9521-498b53b9f2f6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.971671] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "refresh_cache-172647b8-1f1d-49cb-a0fd-63078bc1ae85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.971821] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "refresh_cache-172647b8-1f1d-49cb-a0fd-63078bc1ae85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.971978] env[61985]: DEBUG nova.network.neutron [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.014113] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 964.014113] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 964.052383] env[61985]: DEBUG oslo_concurrency.lockutils [None req-27812e87-cce3-41ea-bd8f-e78a7e3c61c2 tempest-ImagesOneServerTestJSON-2060164899 tempest-ImagesOneServerTestJSON-2060164899-project-member] Lock "518635bf-73ee-404b-ae6a-dc4ee23009d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.503s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.104069] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529b1144-8dbb-bacf-8ba3-cb386f97416d, 'name': SearchDatastore_Task, 'duration_secs': 0.038742} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.104897] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-006d938b-72e9-4eaf-b107-3771afcea9ea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.112487] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for the task: (returnval){ [ 964.112487] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fd6c1d-30f5-e78d-bec4-15a057838492" [ 964.112487] env[61985]: _type = "Task" [ 964.112487] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.120387] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fd6c1d-30f5-e78d-bec4-15a057838492, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.308051] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.356307] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0c905e-6b76-407b-ab15-64df34231f80 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.363977] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749966c8-dbc7-47e1-b7a0-cda92b53c045 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.394383] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9426c0f0-65c6-4455-9ff9-22d02986a4cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.401688] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedf5205-2dc5-4c6b-a60e-d409b7959bb9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.415959] env[61985]: DEBUG nova.compute.provider_tree [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.502304] env[61985]: DEBUG nova.network.neutron [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 964.566659] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.567129] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquired lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.567129] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Forcefully refreshing network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 964.624687] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fd6c1d-30f5-e78d-bec4-15a057838492, 'name': SearchDatastore_Task, 'duration_secs': 0.039204} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.624890] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.625151] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 4da2d122-aa36-4cad-9521-498b53b9f2f6/4da2d122-aa36-4cad-9521-498b53b9f2f6.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 964.625710] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-224b4be0-5727-45b2-bd95-bddb00e75e92 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.632407] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for the task: (returnval){ [ 964.632407] env[61985]: value = "task-936212" [ 964.632407] env[61985]: _type = "Task" [ 964.632407] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.640657] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936212, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.642646] env[61985]: DEBUG nova.compute.manager [req-39601961-0ad2-44d1-ae5c-8205071c5b85 req-50088002-f27d-4790-b9fa-82bc4b5b4bcb service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Received event network-changed-f33869c6-9fa7-44ab-8de2-ae92191b1096 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 964.642885] env[61985]: DEBUG nova.compute.manager [req-39601961-0ad2-44d1-ae5c-8205071c5b85 req-50088002-f27d-4790-b9fa-82bc4b5b4bcb service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Refreshing instance network info cache due to event network-changed-f33869c6-9fa7-44ab-8de2-ae92191b1096. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 964.643094] env[61985]: DEBUG oslo_concurrency.lockutils [req-39601961-0ad2-44d1-ae5c-8205071c5b85 req-50088002-f27d-4790-b9fa-82bc4b5b4bcb service nova] Acquiring lock "refresh_cache-172647b8-1f1d-49cb-a0fd-63078bc1ae85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.709677] env[61985]: DEBUG oslo_concurrency.lockutils [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.710043] env[61985]: DEBUG oslo_concurrency.lockutils [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.710322] env[61985]: DEBUG oslo_concurrency.lockutils [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.710645] env[61985]: DEBUG oslo_concurrency.lockutils [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.710856] env[61985]: DEBUG oslo_concurrency.lockutils [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.713183] env[61985]: INFO nova.compute.manager [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Terminating instance [ 964.715041] env[61985]: DEBUG nova.compute.manager [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 964.715299] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 964.716166] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae864d7a-6e98-4ef9-979b-289bbef0a91d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.723562] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 964.723814] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f26268d-6bf2-43d8-865d-28b5bba530fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.730791] env[61985]: DEBUG oslo_vmware.api [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 964.730791] env[61985]: value = "task-936213" [ 964.730791] env[61985]: _type = "Task" [ 964.730791] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.736253] env[61985]: DEBUG nova.network.neutron [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Updating instance_info_cache with network_info: [{"id": "f33869c6-9fa7-44ab-8de2-ae92191b1096", "address": "fa:16:3e:d1:cd:36", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33869c6-9f", "ovs_interfaceid": "f33869c6-9fa7-44ab-8de2-ae92191b1096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.740494] env[61985]: DEBUG oslo_vmware.api [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936213, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.812231] env[61985]: DEBUG nova.compute.manager [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61985) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 964.812231] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.919317] env[61985]: DEBUG nova.scheduler.client.report [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.142328] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936212, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454317} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.142440] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 4da2d122-aa36-4cad-9521-498b53b9f2f6/4da2d122-aa36-4cad-9521-498b53b9f2f6.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 965.142666] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.142928] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18810f82-7041-4723-ac70-2abdfd67cabb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.149681] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for the task: (returnval){ [ 965.149681] env[61985]: value = "task-936214" [ 965.149681] env[61985]: _type = "Task" [ 965.149681] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.158317] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936214, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.241561] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "refresh_cache-172647b8-1f1d-49cb-a0fd-63078bc1ae85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.241917] env[61985]: DEBUG nova.compute.manager [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Instance network_info: |[{"id": "f33869c6-9fa7-44ab-8de2-ae92191b1096", "address": "fa:16:3e:d1:cd:36", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33869c6-9f", "ovs_interfaceid": "f33869c6-9fa7-44ab-8de2-ae92191b1096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.242263] env[61985]: DEBUG oslo_vmware.api [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936213, 'name': PowerOffVM_Task, 'duration_secs': 0.356785} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.242503] env[61985]: DEBUG oslo_concurrency.lockutils [req-39601961-0ad2-44d1-ae5c-8205071c5b85 req-50088002-f27d-4790-b9fa-82bc4b5b4bcb service nova] Acquired lock "refresh_cache-172647b8-1f1d-49cb-a0fd-63078bc1ae85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.242688] env[61985]: DEBUG nova.network.neutron [req-39601961-0ad2-44d1-ae5c-8205071c5b85 req-50088002-f27d-4790-b9fa-82bc4b5b4bcb service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Refreshing network info cache for port f33869c6-9fa7-44ab-8de2-ae92191b1096 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 965.244096] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:cd:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd829efb7-e98e-4b67-bd03-b0888287dbfd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f33869c6-9fa7-44ab-8de2-ae92191b1096', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.253100] env[61985]: DEBUG oslo.service.loopingcall [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.253622] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 965.253905] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 965.257154] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 965.257532] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c968d3cc-eb60-4c0c-a44f-21bb1d3c416a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.259899] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4d2ce830-e5a8-4619-af2c-0940a633a4e9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.281336] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.281336] env[61985]: value = "task-936216" [ 965.281336] env[61985]: _type = "Task" [ 965.281336] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.292133] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936216, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.344514] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 965.345018] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 965.345285] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleting the datastore file [datastore1] 57591f6f-ed7d-46f3-a9c1-217016c42b1c {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.345619] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25efc9d1-927c-44ef-8a82-d88c233aebf7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.351996] env[61985]: DEBUG oslo_vmware.api [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 965.351996] env[61985]: value = "task-936217" [ 965.351996] env[61985]: _type = "Task" [ 965.351996] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.362349] env[61985]: DEBUG oslo_vmware.api [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936217, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.425546] env[61985]: DEBUG oslo_concurrency.lockutils [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.427672] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.477s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.427948] env[61985]: DEBUG nova.objects.instance [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lazy-loading 'resources' on Instance uuid 62824eac-9412-466a-abcf-1010f6a829e4 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.456199] env[61985]: INFO nova.scheduler.client.report [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleted allocations for instance cd8e2eaf-3c13-452b-b2e6-8107f0219378 [ 965.660784] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936214, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109057} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.661528] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.662456] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d93b06-1007-4597-9c6b-fecef39e47c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.690920] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 4da2d122-aa36-4cad-9521-498b53b9f2f6/4da2d122-aa36-4cad-9521-498b53b9f2f6.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.691349] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e2a2cd5-e6eb-435b-b878-1fb65fc9d4bf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.715094] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for the task: (returnval){ [ 965.715094] env[61985]: value = "task-936218" [ 965.715094] env[61985]: _type = "Task" [ 965.715094] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.729330] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936218, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.731292] env[61985]: DEBUG nova.network.neutron [req-39601961-0ad2-44d1-ae5c-8205071c5b85 req-50088002-f27d-4790-b9fa-82bc4b5b4bcb service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Updated VIF entry in instance network info cache for port f33869c6-9fa7-44ab-8de2-ae92191b1096. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 965.731651] env[61985]: DEBUG nova.network.neutron [req-39601961-0ad2-44d1-ae5c-8205071c5b85 req-50088002-f27d-4790-b9fa-82bc4b5b4bcb service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Updating instance_info_cache with network_info: [{"id": "f33869c6-9fa7-44ab-8de2-ae92191b1096", "address": "fa:16:3e:d1:cd:36", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33869c6-9f", "ovs_interfaceid": "f33869c6-9fa7-44ab-8de2-ae92191b1096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.792254] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936216, 'name': CreateVM_Task, 'duration_secs': 0.363029} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.792254] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 965.792254] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.792254] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.792501] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 965.795105] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9210a3b-76a4-4aa0-a526-c8b63ea21c3c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.800029] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 965.800029] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52da4ed6-5e68-939d-b603-ce494a400795" [ 965.800029] env[61985]: _type = "Task" [ 965.800029] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.809284] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52da4ed6-5e68-939d-b603-ce494a400795, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.862507] env[61985]: DEBUG oslo_vmware.api [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936217, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137479} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.864712] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.864712] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 965.864712] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 965.864712] env[61985]: INFO nova.compute.manager [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 965.864712] env[61985]: DEBUG oslo.service.loopingcall [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.864712] env[61985]: DEBUG nova.compute.manager [-] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.864712] env[61985]: DEBUG nova.network.neutron [-] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 965.969318] env[61985]: DEBUG oslo_concurrency.lockutils [None req-881116fd-1f03-4803-96a9-3050f9a3671d tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "cd8e2eaf-3c13-452b-b2e6-8107f0219378" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.366s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.059199] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance_info_cache with network_info: [{"id": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "address": "fa:16:3e:7b:a1:a8", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.137", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadde3d48-eb", "ovs_interfaceid": "adde3d48-ebd4-450d-920a-fb93eeff4a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.235122] env[61985]: DEBUG oslo_concurrency.lockutils [req-39601961-0ad2-44d1-ae5c-8205071c5b85 req-50088002-f27d-4790-b9fa-82bc4b5b4bcb service nova] Releasing lock "refresh_cache-172647b8-1f1d-49cb-a0fd-63078bc1ae85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.235622] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936218, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.312651] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52da4ed6-5e68-939d-b603-ce494a400795, 'name': SearchDatastore_Task, 'duration_secs': 0.009425} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.313092] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.313419] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.313727] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.313944] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.314202] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.314550] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e51bb596-b06a-429c-b101-26ac97db7e01 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.323614] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.323896] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 966.324702] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3097001-5bb9-40b7-932a-ad4420b14f27 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.330631] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 966.330631] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5257c61c-b5c4-1300-c95e-5c47d81dcc1f" [ 966.330631] env[61985]: _type = "Task" [ 966.330631] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.339496] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5257c61c-b5c4-1300-c95e-5c47d81dcc1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.434580] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010b4404-fbd1-4f3a-8b73-a944397ef799 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.442347] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db09b23-27da-4b8c-9edc-7da69b44e9cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.472468] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9083ddf-2c21-4171-8c9b-c9d68d33806f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.480006] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40020785-c215-4788-8603-cfe193232d5c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.493080] env[61985]: DEBUG nova.compute.provider_tree [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.562777] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Releasing lock "refresh_cache-fea0cf39-e851-409f-86f5-31cc128a44dc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.563032] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updated the network info_cache for instance {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10003}} [ 966.563260] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.563416] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.563648] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.563725] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.563837] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.563973] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.564120] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 966.564274] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.667106] env[61985]: DEBUG nova.compute.manager [req-398666e9-16aa-49f6-98e5-5f9d0590a120 req-34d18014-a4e2-4745-a502-d937c5ff56fb service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Received event network-vif-deleted-2ace42f2-9857-4a83-bcd2-75e3f0a3e788 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 966.667341] env[61985]: INFO nova.compute.manager [req-398666e9-16aa-49f6-98e5-5f9d0590a120 req-34d18014-a4e2-4745-a502-d937c5ff56fb service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Neutron deleted interface 2ace42f2-9857-4a83-bcd2-75e3f0a3e788; detaching it from the instance and deleting it from the info cache [ 966.667434] env[61985]: DEBUG nova.network.neutron [req-398666e9-16aa-49f6-98e5-5f9d0590a120 req-34d18014-a4e2-4745-a502-d937c5ff56fb service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.711539] env[61985]: DEBUG nova.network.neutron [-] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.728507] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936218, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.842049] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5257c61c-b5c4-1300-c95e-5c47d81dcc1f, 'name': SearchDatastore_Task, 'duration_secs': 0.013739} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.842487] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0010fa2-5bf5-49a3-917d-c77837978702 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.847922] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 966.847922] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ba1a29-9e32-3703-8c51-ef3344d16e27" [ 966.847922] env[61985]: _type = "Task" [ 966.847922] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.856946] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ba1a29-9e32-3703-8c51-ef3344d16e27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.995910] env[61985]: DEBUG nova.scheduler.client.report [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.068068] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.169808] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6ce073b-601f-4f70-a5f8-8a47a405f4b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.179358] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b3879d-5f3a-4efc-9596-a9bcc3cb2042 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.211902] env[61985]: DEBUG nova.compute.manager [req-398666e9-16aa-49f6-98e5-5f9d0590a120 req-34d18014-a4e2-4745-a502-d937c5ff56fb service nova] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Detach interface failed, port_id=2ace42f2-9857-4a83-bcd2-75e3f0a3e788, reason: Instance 57591f6f-ed7d-46f3-a9c1-217016c42b1c could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 967.215620] env[61985]: INFO nova.compute.manager [-] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Took 1.35 seconds to deallocate network for instance. [ 967.228926] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936218, 'name': ReconfigVM_Task, 'duration_secs': 1.1431} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.229403] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 4da2d122-aa36-4cad-9521-498b53b9f2f6/4da2d122-aa36-4cad-9521-498b53b9f2f6.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.230745] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4dd0d38-56e0-46c5-b0e3-988638e4b7d8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.238283] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for the task: (returnval){ [ 967.238283] env[61985]: value = "task-936219" [ 967.238283] env[61985]: _type = "Task" [ 967.238283] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.245728] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936219, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.361017] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ba1a29-9e32-3703-8c51-ef3344d16e27, 'name': SearchDatastore_Task, 'duration_secs': 0.009806} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.361017] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.361017] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 172647b8-1f1d-49cb-a0fd-63078bc1ae85/172647b8-1f1d-49cb-a0fd-63078bc1ae85.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 967.361017] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-353e3a4d-289c-4d2e-9a8c-2ebf2bfc5677 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.366344] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 967.366344] env[61985]: value = "task-936220" [ 967.366344] env[61985]: _type = "Task" [ 967.366344] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.375115] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936220, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.500999] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.073s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.503394] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.522s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.503634] env[61985]: DEBUG nova.objects.instance [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lazy-loading 'resources' on Instance uuid 1f284789-1e7b-4e9f-9670-34e8e25cd797 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.529956] env[61985]: INFO nova.scheduler.client.report [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Deleted allocations for instance 62824eac-9412-466a-abcf-1010f6a829e4 [ 967.725783] env[61985]: DEBUG oslo_concurrency.lockutils [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.751412] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936219, 'name': Rename_Task, 'duration_secs': 0.193473} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.751706] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 967.751973] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbdebc02-2e42-46ed-9d28-a8a61b37fd22 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.759973] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for the task: (returnval){ [ 967.759973] env[61985]: value = "task-936221" [ 967.759973] env[61985]: _type = "Task" [ 967.759973] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.769111] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936221, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.880988] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936220, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.042885] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3b0a293b-384c-400d-8207-e99b7709f713 tempest-MultipleCreateTestJSON-1576053803 tempest-MultipleCreateTestJSON-1576053803-project-member] Lock "62824eac-9412-466a-abcf-1010f6a829e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.268s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.272811] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936221, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.383242] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936220, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579808} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.383611] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 172647b8-1f1d-49cb-a0fd-63078bc1ae85/172647b8-1f1d-49cb-a0fd-63078bc1ae85.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 968.387021] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.387021] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4be11158-c10c-4017-840f-b7b0fa60ac0f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.392027] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600d13d6-f936-45de-a9cb-0d8bdf96e462 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.396230] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 968.396230] env[61985]: value = "task-936222" [ 968.396230] env[61985]: _type = "Task" [ 968.396230] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.405084] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45128ca4-800e-4ee1-ab6f-d8f2cb144c8c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.412736] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936222, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.442013] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8671e4-4f05-43aa-acde-839a642d3ef3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.450020] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b51f4d4-3be4-4390-ae08-7642897c294d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.466099] env[61985]: DEBUG nova.compute.provider_tree [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.778759] env[61985]: DEBUG oslo_vmware.api [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936221, 'name': PowerOnVM_Task, 'duration_secs': 0.625741} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.779706] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 968.780276] env[61985]: INFO nova.compute.manager [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Took 9.77 seconds to spawn the instance on the hypervisor. [ 968.781148] env[61985]: DEBUG nova.compute.manager [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.788164] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08982b15-f335-4e64-afea-e2e81ddd7d88 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.910855] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936222, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071023} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.911381] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.912319] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f72473b-9e48-4484-9f8e-53dab7f29587 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.936176] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 172647b8-1f1d-49cb-a0fd-63078bc1ae85/172647b8-1f1d-49cb-a0fd-63078bc1ae85.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.936703] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0250e78-4657-4a00-b6eb-9b57d877affa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.958556] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 968.958556] env[61985]: value = "task-936223" [ 968.958556] env[61985]: _type = "Task" [ 968.958556] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.966582] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936223, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.968609] env[61985]: DEBUG nova.scheduler.client.report [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.306056] env[61985]: INFO nova.compute.manager [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Took 34.83 seconds to build instance. [ 969.475230] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936223, 'name': ReconfigVM_Task, 'duration_secs': 0.314623} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.476114] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.973s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.479048] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 172647b8-1f1d-49cb-a0fd-63078bc1ae85/172647b8-1f1d-49cb-a0fd-63078bc1ae85.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.482233] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.274s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.482233] env[61985]: INFO nova.compute.claims [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.484893] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a9bd9447-f5d9-48c3-ae02-d2dee77e3626 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.495627] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 969.495627] env[61985]: value = "task-936224" [ 969.495627] env[61985]: _type = "Task" [ 969.495627] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.508046] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936224, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.516105] env[61985]: INFO nova.scheduler.client.report [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Deleted allocations for instance 1f284789-1e7b-4e9f-9670-34e8e25cd797 [ 969.809781] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f958fe8e-912a-485f-ba60-a6a13ba56748 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lock "4da2d122-aa36-4cad-9521-498b53b9f2f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.343s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.018014] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936224, 'name': Rename_Task, 'duration_secs': 0.138139} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.018461] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 970.018854] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-090dee41-3fb9-4866-a21a-5219365cf2d5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.031765] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0afa4c-b4e8-413f-8141-617bdddbf867 tempest-AttachVolumeTestJSON-482174914 tempest-AttachVolumeTestJSON-482174914-project-member] Lock "1f284789-1e7b-4e9f-9670-34e8e25cd797" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.897s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.033142] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 970.033142] env[61985]: value = "task-936225" [ 970.033142] env[61985]: _type = "Task" [ 970.033142] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.041464] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936225, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.547685] env[61985]: DEBUG oslo_vmware.api [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936225, 'name': PowerOnVM_Task, 'duration_secs': 0.447562} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.548124] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 970.548216] env[61985]: INFO nova.compute.manager [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Took 9.05 seconds to spawn the instance on the hypervisor. [ 970.548464] env[61985]: DEBUG nova.compute.manager [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.549849] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9ae419-dbf9-409e-b664-399cdbe52054 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.601708] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquiring lock "bed26b5f-7ca5-405a-884e-02b2495dd977" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.601708] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lock "bed26b5f-7ca5-405a-884e-02b2495dd977" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.902594] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e0ff70-c8ec-464e-9cc2-62b4bd618f23 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.914992] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a97897-a9f8-42d8-aa4d-62154765c5be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.960176] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1899d9e4-c576-48a1-a103-6526a6de7d6d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.970482] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120e5624-d6ac-4561-9913-47b5bb140dab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.987021] env[61985]: DEBUG nova.compute.provider_tree [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.079271] env[61985]: INFO nova.compute.manager [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Took 35.86 seconds to build instance. [ 971.105826] env[61985]: DEBUG nova.compute.manager [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.491116] env[61985]: DEBUG nova.scheduler.client.report [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.582956] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9e6cdba8-6361-4811-9334-37bb55c70b82 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.379s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.634425] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.758941] env[61985]: DEBUG nova.compute.manager [req-532ff6b8-c5af-4765-8cea-bfb281e9a862 req-f52f8521-6d1c-4f66-8321-f89d71c8c665 service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Received event network-changed-f33869c6-9fa7-44ab-8de2-ae92191b1096 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 971.760043] env[61985]: DEBUG nova.compute.manager [req-532ff6b8-c5af-4765-8cea-bfb281e9a862 req-f52f8521-6d1c-4f66-8321-f89d71c8c665 service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Refreshing instance network info cache due to event network-changed-f33869c6-9fa7-44ab-8de2-ae92191b1096. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 971.760428] env[61985]: DEBUG oslo_concurrency.lockutils [req-532ff6b8-c5af-4765-8cea-bfb281e9a862 req-f52f8521-6d1c-4f66-8321-f89d71c8c665 service nova] Acquiring lock "refresh_cache-172647b8-1f1d-49cb-a0fd-63078bc1ae85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.760595] env[61985]: DEBUG oslo_concurrency.lockutils [req-532ff6b8-c5af-4765-8cea-bfb281e9a862 req-f52f8521-6d1c-4f66-8321-f89d71c8c665 service nova] Acquired lock "refresh_cache-172647b8-1f1d-49cb-a0fd-63078bc1ae85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.760762] env[61985]: DEBUG nova.network.neutron [req-532ff6b8-c5af-4765-8cea-bfb281e9a862 req-f52f8521-6d1c-4f66-8321-f89d71c8c665 service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Refreshing network info cache for port f33869c6-9fa7-44ab-8de2-ae92191b1096 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 971.996443] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.996978] env[61985]: DEBUG nova.compute.manager [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 972.000858] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.576s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.002311] env[61985]: INFO nova.compute.claims [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.418627] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquiring lock "4da2d122-aa36-4cad-9521-498b53b9f2f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.418944] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lock "4da2d122-aa36-4cad-9521-498b53b9f2f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.419146] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquiring lock "4da2d122-aa36-4cad-9521-498b53b9f2f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.419336] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lock "4da2d122-aa36-4cad-9521-498b53b9f2f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.419507] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lock "4da2d122-aa36-4cad-9521-498b53b9f2f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.422068] env[61985]: INFO nova.compute.manager [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Terminating instance [ 972.580360] env[61985]: DEBUG nova.compute.manager [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 972.580360] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 972.580360] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d06eeb7-14c1-460c-b8c9-bd0ef68ca3a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.580360] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 972.580360] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67f92ce9-6a55-4a54-8901-f442e456d8b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.580360] env[61985]: DEBUG oslo_vmware.api [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for the task: (returnval){ [ 972.580360] env[61985]: value = "task-936227" [ 972.580360] env[61985]: _type = "Task" [ 972.580360] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.580360] env[61985]: DEBUG oslo_vmware.api [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936227, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.580360] env[61985]: DEBUG nova.compute.utils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 972.580360] env[61985]: DEBUG nova.compute.manager [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 972.580360] env[61985]: DEBUG nova.network.neutron [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 972.589393] env[61985]: DEBUG nova.policy [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da7c7d5d10b34b4ebdede5ff3cfd8b01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e66b7a6e7354b06a77295eadf4e5e0f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 972.672151] env[61985]: DEBUG nova.network.neutron [req-532ff6b8-c5af-4765-8cea-bfb281e9a862 req-f52f8521-6d1c-4f66-8321-f89d71c8c665 service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Updated VIF entry in instance network info cache for port f33869c6-9fa7-44ab-8de2-ae92191b1096. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 972.672499] env[61985]: DEBUG nova.network.neutron [req-532ff6b8-c5af-4765-8cea-bfb281e9a862 req-f52f8521-6d1c-4f66-8321-f89d71c8c665 service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Updating instance_info_cache with network_info: [{"id": "f33869c6-9fa7-44ab-8de2-ae92191b1096", "address": "fa:16:3e:d1:cd:36", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33869c6-9f", "ovs_interfaceid": "f33869c6-9fa7-44ab-8de2-ae92191b1096", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.986136] env[61985]: DEBUG oslo_vmware.api [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936227, 'name': PowerOffVM_Task, 'duration_secs': 0.186549} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.986136] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 972.986136] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 972.986136] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-87c3970f-0e7b-4117-8c80-d1baa4e74b8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.013901] env[61985]: DEBUG nova.network.neutron [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Successfully created port: c7cba622-8372-4f2f-994a-8c92dab641bb {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 973.024727] env[61985]: DEBUG nova.compute.manager [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 973.052692] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 973.053086] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 973.053647] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Deleting the datastore file [datastore1] 4da2d122-aa36-4cad-9521-498b53b9f2f6 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.053647] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-673be7ce-f43c-4846-886b-105c94fad9b1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.060169] env[61985]: DEBUG oslo_vmware.api [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for the task: (returnval){ [ 973.060169] env[61985]: value = "task-936229" [ 973.060169] env[61985]: _type = "Task" [ 973.060169] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.074159] env[61985]: DEBUG oslo_vmware.api [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936229, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.178211] env[61985]: DEBUG oslo_concurrency.lockutils [req-532ff6b8-c5af-4765-8cea-bfb281e9a862 req-f52f8521-6d1c-4f66-8321-f89d71c8c665 service nova] Releasing lock "refresh_cache-172647b8-1f1d-49cb-a0fd-63078bc1ae85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.435531] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05314c2f-960d-49dd-90ac-a2e56b6d418c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.445144] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf58d5d-48b5-4026-ada2-806264a1fe90 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.490608] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a159ba1b-ae51-486a-8e01-05e703be47a0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.499122] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd1ebc0-2a3f-4788-8b66-85aab56514d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.513052] env[61985]: DEBUG nova.compute.provider_tree [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.570832] env[61985]: DEBUG oslo_vmware.api [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Task: {'id': task-936229, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205933} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.571125] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.571312] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 973.571495] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 973.571680] env[61985]: INFO nova.compute.manager [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Took 1.14 seconds to destroy the instance on the hypervisor. [ 973.571932] env[61985]: DEBUG oslo.service.loopingcall [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.572156] env[61985]: DEBUG nova.compute.manager [-] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.572250] env[61985]: DEBUG nova.network.neutron [-] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 973.906358] env[61985]: DEBUG nova.compute.manager [req-fa789612-e2f9-4837-83d7-29d1b441e6b9 req-56c6713c-ed6d-4cb0-ab6f-1b877793306a service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Received event network-vif-deleted-92893cd3-94c2-4842-a021-372bd8d06fda {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 973.906358] env[61985]: INFO nova.compute.manager [req-fa789612-e2f9-4837-83d7-29d1b441e6b9 req-56c6713c-ed6d-4cb0-ab6f-1b877793306a service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Neutron deleted interface 92893cd3-94c2-4842-a021-372bd8d06fda; detaching it from the instance and deleting it from the info cache [ 973.906358] env[61985]: DEBUG nova.network.neutron [req-fa789612-e2f9-4837-83d7-29d1b441e6b9 req-56c6713c-ed6d-4cb0-ab6f-1b877793306a service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.015597] env[61985]: DEBUG nova.scheduler.client.report [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.037306] env[61985]: DEBUG nova.compute.manager [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 974.074572] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='55d17a48046daaba700386878f734317',container_format='bare',created_at=2024-09-18T01:30:17Z,direct_url=,disk_format='vmdk',id=b4b4e229-a7cf-45de-8912-47f4b206140b,min_disk=1,min_ram=0,name='tempest-test-snap-789495471',owner='2e66b7a6e7354b06a77295eadf4e5e0f',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-09-18T01:30:35Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.074832] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.074998] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.076133] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.076133] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.076266] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.076386] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.076555] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.076730] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.076898] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.077096] env[61985]: DEBUG nova.virt.hardware [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.078283] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c73f4b-1b44-4d44-960f-4509e15b55b6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.086568] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afc8fd1-1544-4c7e-af05-2dfe9ea5d8ad {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.344852] env[61985]: DEBUG nova.network.neutron [-] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.416128] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3453e8b-523b-4162-9523-2bff05b40339 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.425356] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843e357f-b4cb-43a7-aae3-f63076533915 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.463306] env[61985]: DEBUG nova.compute.manager [req-fa789612-e2f9-4837-83d7-29d1b441e6b9 req-56c6713c-ed6d-4cb0-ab6f-1b877793306a service nova] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Detach interface failed, port_id=92893cd3-94c2-4842-a021-372bd8d06fda, reason: Instance 4da2d122-aa36-4cad-9521-498b53b9f2f6 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 974.524328] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.524328] env[61985]: DEBUG nova.compute.manager [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 974.525945] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 23.438s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.850278] env[61985]: INFO nova.compute.manager [-] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Took 1.28 seconds to deallocate network for instance. [ 974.986355] env[61985]: DEBUG nova.network.neutron [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Successfully updated port: c7cba622-8372-4f2f-994a-8c92dab641bb {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.030217] env[61985]: DEBUG nova.compute.utils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.035367] env[61985]: DEBUG nova.compute.manager [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 975.035367] env[61985]: DEBUG nova.network.neutron [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 975.122866] env[61985]: DEBUG nova.policy [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a78dbc667d24c819e6447740fc2d8e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2dbceebf07c43ac90ed80f8a09294b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 975.358046] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.437080] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764394a5-13cf-475d-be91-970631f45a95 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.445386] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc0b959-7c58-4238-892c-509d7071b817 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.476982] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a1cf89-65f3-4dc4-b81c-ed26e7a550dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.484983] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c92d479-e77c-4036-9a4c-1a2a32a8ab7e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.489549] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "refresh_cache-942e0f09-4d34-4aa2-8a60-4c12f899712c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.489694] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "refresh_cache-942e0f09-4d34-4aa2-8a60-4c12f899712c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.489851] env[61985]: DEBUG nova.network.neutron [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 975.507774] env[61985]: DEBUG nova.compute.provider_tree [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.537028] env[61985]: DEBUG nova.compute.manager [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 975.657475] env[61985]: DEBUG nova.network.neutron [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Successfully created port: 7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.876906] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "44361f7b-4609-476f-b4a9-58a7851e6e92" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.877182] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.931395] env[61985]: DEBUG nova.compute.manager [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Received event network-vif-plugged-c7cba622-8372-4f2f-994a-8c92dab641bb {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 975.931634] env[61985]: DEBUG oslo_concurrency.lockutils [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] Acquiring lock "942e0f09-4d34-4aa2-8a60-4c12f899712c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.931942] env[61985]: DEBUG oslo_concurrency.lockutils [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] Lock "942e0f09-4d34-4aa2-8a60-4c12f899712c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.932025] env[61985]: DEBUG oslo_concurrency.lockutils [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] Lock "942e0f09-4d34-4aa2-8a60-4c12f899712c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.932203] env[61985]: DEBUG nova.compute.manager [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] No waiting events found dispatching network-vif-plugged-c7cba622-8372-4f2f-994a-8c92dab641bb {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 975.932375] env[61985]: WARNING nova.compute.manager [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Received unexpected event network-vif-plugged-c7cba622-8372-4f2f-994a-8c92dab641bb for instance with vm_state building and task_state spawning. [ 975.932545] env[61985]: DEBUG nova.compute.manager [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Received event network-changed-c7cba622-8372-4f2f-994a-8c92dab641bb {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 975.932830] env[61985]: DEBUG nova.compute.manager [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Refreshing instance network info cache due to event network-changed-c7cba622-8372-4f2f-994a-8c92dab641bb. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 975.932888] env[61985]: DEBUG oslo_concurrency.lockutils [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] Acquiring lock "refresh_cache-942e0f09-4d34-4aa2-8a60-4c12f899712c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.012676] env[61985]: DEBUG nova.scheduler.client.report [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.047234] env[61985]: DEBUG nova.network.neutron [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 976.245433] env[61985]: DEBUG nova.network.neutron [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Updating instance_info_cache with network_info: [{"id": "c7cba622-8372-4f2f-994a-8c92dab641bb", "address": "fa:16:3e:4f:74:4c", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7cba622-83", "ovs_interfaceid": "c7cba622-8372-4f2f-994a-8c92dab641bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.380459] env[61985]: DEBUG nova.compute.manager [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 976.548162] env[61985]: DEBUG nova.compute.manager [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 976.583942] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.584445] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.584694] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.585195] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.586545] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.586545] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.586545] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.586545] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.586545] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.586545] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.586545] env[61985]: DEBUG nova.virt.hardware [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.587410] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f5511c-b131-4941-8256-6d537e34c054 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.597177] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be067ef3-8b4a-47f4-b83e-f8cc76a5e8bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.749867] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "refresh_cache-942e0f09-4d34-4aa2-8a60-4c12f899712c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.750584] env[61985]: DEBUG nova.compute.manager [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Instance network_info: |[{"id": "c7cba622-8372-4f2f-994a-8c92dab641bb", "address": "fa:16:3e:4f:74:4c", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7cba622-83", "ovs_interfaceid": "c7cba622-8372-4f2f-994a-8c92dab641bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 976.751097] env[61985]: DEBUG oslo_concurrency.lockutils [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] Acquired lock "refresh_cache-942e0f09-4d34-4aa2-8a60-4c12f899712c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.751331] env[61985]: DEBUG nova.network.neutron [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Refreshing network info cache for port c7cba622-8372-4f2f-994a-8c92dab641bb {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 976.752555] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:74:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2c019b6-3ef3-4c8f-95bd-edede2c554a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7cba622-8372-4f2f-994a-8c92dab641bb', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 976.760640] env[61985]: DEBUG oslo.service.loopingcall [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.761195] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 976.761466] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-796af3f4-0cf1-4df4-bea2-e34b14292949 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.783652] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 976.783652] env[61985]: value = "task-936231" [ 976.783652] env[61985]: _type = "Task" [ 976.783652] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.791910] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936231, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.905493] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.023618] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.498s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.023912] env[61985]: DEBUG nova.compute.manager [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61985) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4915}} [ 977.030410] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.418s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.030688] env[61985]: DEBUG nova.objects.instance [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lazy-loading 'resources' on Instance uuid 488d0b95-97a9-4193-af62-cc92caf99625 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.299019] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936231, 'name': CreateVM_Task, 'duration_secs': 0.358625} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.299019] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 977.299019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.299019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.299019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.299019] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fd6b130-b90e-401c-8d90-9f395651a81e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.302865] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 977.302865] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b47236-a94f-be19-5a52-dfcfda5af1b7" [ 977.302865] env[61985]: _type = "Task" [ 977.302865] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.312865] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b47236-a94f-be19-5a52-dfcfda5af1b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.411493] env[61985]: DEBUG nova.network.neutron [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Successfully updated port: 7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.624242] env[61985]: INFO nova.scheduler.client.report [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleted allocation for migration 9a221221-2558-443c-8a47-3d0ca57efb01 [ 977.661918] env[61985]: DEBUG nova.network.neutron [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Updated VIF entry in instance network info cache for port c7cba622-8372-4f2f-994a-8c92dab641bb. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 977.662378] env[61985]: DEBUG nova.network.neutron [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Updating instance_info_cache with network_info: [{"id": "c7cba622-8372-4f2f-994a-8c92dab641bb", "address": "fa:16:3e:4f:74:4c", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7cba622-83", "ovs_interfaceid": "c7cba622-8372-4f2f-994a-8c92dab641bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.821616] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.821923] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Processing image b4b4e229-a7cf-45de-8912-47f4b206140b {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.822337] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b/b4b4e229-a7cf-45de-8912-47f4b206140b.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.822507] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b/b4b4e229-a7cf-45de-8912-47f4b206140b.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.822709] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.823617] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-86898a58-6d8b-4751-bd7c-b088a79660bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.837800] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.838081] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 977.838766] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91be1eab-9f19-4b87-acf6-026850cebc0c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.845013] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 977.845013] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529c3c3a-c57a-ce20-60bd-d620ea12e0b8" [ 977.845013] env[61985]: _type = "Task" [ 977.845013] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.854839] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529c3c3a-c57a-ce20-60bd-d620ea12e0b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.913709] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.914309] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.914309] env[61985]: DEBUG nova.network.neutron [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.948898] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07bee55a-f74c-47dd-b144-11e89d34f268 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.958148] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc739c07-4074-4838-b4a1-217d1d334a3f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.987382] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7df28f-1b1e-46cf-be25-3a84ed1b9fe2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.995465] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5207cf-31c3-4a03-af63-bcec70ccc9ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.010334] env[61985]: DEBUG nova.compute.provider_tree [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.134060] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8421f1c4-b76d-47fa-9798-d88f81edf9f4 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 30.655s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.165339] env[61985]: DEBUG oslo_concurrency.lockutils [req-ece5f683-8ea7-4286-a6e8-9c3e2508f752 req-c57c3da9-6f29-4a2d-aa2f-3ef73941c0b8 service nova] Releasing lock "refresh_cache-942e0f09-4d34-4aa2-8a60-4c12f899712c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.316399] env[61985]: DEBUG nova.objects.instance [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'flavor' on Instance uuid 2322ab72-9841-41fb-9d60-2812baabe108 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.354883] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Preparing fetch location {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 978.355173] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Fetch image to [datastore2] OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda/OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda.vmdk {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 978.355397] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Downloading stream optimized image b4b4e229-a7cf-45de-8912-47f4b206140b to [datastore2] OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda/OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda.vmdk on the data store datastore2 as vApp {{(pid=61985) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 978.355539] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Downloading image file data b4b4e229-a7cf-45de-8912-47f4b206140b to the ESX as VM named 'OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda' {{(pid=61985) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 978.437732] env[61985]: DEBUG nova.compute.manager [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received event network-vif-plugged-7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 978.437993] env[61985]: DEBUG oslo_concurrency.lockutils [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] Acquiring lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.438227] env[61985]: DEBUG oslo_concurrency.lockutils [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] Lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.438426] env[61985]: DEBUG oslo_concurrency.lockutils [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] Lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.438581] env[61985]: DEBUG nova.compute.manager [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] No waiting events found dispatching network-vif-plugged-7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 978.438753] env[61985]: WARNING nova.compute.manager [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received unexpected event network-vif-plugged-7075efce-177a-4127-9c2f-7e3a9d1e92ad for instance with vm_state building and task_state spawning. [ 978.438921] env[61985]: DEBUG nova.compute.manager [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received event network-changed-7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 978.439096] env[61985]: DEBUG nova.compute.manager [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing instance network info cache due to event network-changed-7075efce-177a-4127-9c2f-7e3a9d1e92ad. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 978.439273] env[61985]: DEBUG oslo_concurrency.lockutils [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] Acquiring lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.443891] env[61985]: DEBUG oslo_vmware.rw_handles [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 978.443891] env[61985]: value = "resgroup-9" [ 978.443891] env[61985]: _type = "ResourcePool" [ 978.443891] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 978.444231] env[61985]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-e26b197f-e4ff-44dc-91f6-f0d3fbaa434c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.460698] env[61985]: DEBUG nova.network.neutron [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 978.470277] env[61985]: DEBUG oslo_vmware.rw_handles [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lease: (returnval){ [ 978.470277] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e86d5b-b1cd-ff3f-006d-50187acdbfe5" [ 978.470277] env[61985]: _type = "HttpNfcLease" [ 978.470277] env[61985]: } obtained for vApp import into resource pool (val){ [ 978.470277] env[61985]: value = "resgroup-9" [ 978.470277] env[61985]: _type = "ResourcePool" [ 978.470277] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 978.470594] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the lease: (returnval){ [ 978.470594] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e86d5b-b1cd-ff3f-006d-50187acdbfe5" [ 978.470594] env[61985]: _type = "HttpNfcLease" [ 978.470594] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 978.477424] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 978.477424] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e86d5b-b1cd-ff3f-006d-50187acdbfe5" [ 978.477424] env[61985]: _type = "HttpNfcLease" [ 978.477424] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 978.513804] env[61985]: DEBUG nova.scheduler.client.report [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.759593] env[61985]: DEBUG nova.network.neutron [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [{"id": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "address": "fa:16:3e:4b:3d:56", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7075efce-17", "ovs_interfaceid": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.821781] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.821985] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.822187] env[61985]: DEBUG nova.network.neutron [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 978.822387] env[61985]: DEBUG nova.objects.instance [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'info_cache' on Instance uuid 2322ab72-9841-41fb-9d60-2812baabe108 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.978563] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 978.978563] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e86d5b-b1cd-ff3f-006d-50187acdbfe5" [ 978.978563] env[61985]: _type = "HttpNfcLease" [ 978.978563] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 978.978931] env[61985]: DEBUG oslo_vmware.rw_handles [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 978.978931] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e86d5b-b1cd-ff3f-006d-50187acdbfe5" [ 978.978931] env[61985]: _type = "HttpNfcLease" [ 978.978931] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 978.979648] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4fd894f-0866-4e4c-aaa6-54fa255371db {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.986456] env[61985]: DEBUG oslo_vmware.rw_handles [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52948eac-c8bd-df4c-2c95-6a8663865dba/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 978.986640] env[61985]: DEBUG oslo_vmware.rw_handles [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52948eac-c8bd-df4c-2c95-6a8663865dba/disk-0.vmdk. {{(pid=61985) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 979.043989] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.013s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.046036] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 24.211s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.052710] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d4295cb0-1993-4a65-93f4-826673526872 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.067792] env[61985]: INFO nova.scheduler.client.report [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Deleted allocations for instance 488d0b95-97a9-4193-af62-cc92caf99625 [ 979.262471] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.262895] env[61985]: DEBUG nova.compute.manager [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Instance network_info: |[{"id": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "address": "fa:16:3e:4b:3d:56", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7075efce-17", "ovs_interfaceid": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 979.263340] env[61985]: DEBUG oslo_concurrency.lockutils [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] Acquired lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.263623] env[61985]: DEBUG nova.network.neutron [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing network info cache for port 7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 979.264931] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:3d:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7075efce-177a-4127-9c2f-7e3a9d1e92ad', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.274032] env[61985]: DEBUG oslo.service.loopingcall [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.283472] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 979.285352] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ad2afc4-78a8-4e3f-b314-34525940561c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.312411] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 979.312411] env[61985]: value = "task-936233" [ 979.312411] env[61985]: _type = "Task" [ 979.312411] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.323755] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936233, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.326542] env[61985]: DEBUG nova.objects.base [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Object Instance<2322ab72-9841-41fb-9d60-2812baabe108> lazy-loaded attributes: flavor,info_cache {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 979.550891] env[61985]: INFO nova.compute.claims [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.579139] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c25690c0-f84f-4fbe-9c7d-e90fe21d1756 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "488d0b95-97a9-4193-af62-cc92caf99625" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.239s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.753017] env[61985]: DEBUG nova.network.neutron [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updated VIF entry in instance network info cache for port 7075efce-177a-4127-9c2f-7e3a9d1e92ad. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 979.753017] env[61985]: DEBUG nova.network.neutron [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [{"id": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "address": "fa:16:3e:4b:3d:56", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7075efce-17", "ovs_interfaceid": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.825026] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936233, 'name': CreateVM_Task, 'duration_secs': 0.455265} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.825026] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 979.825026] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.825026] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.825026] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.825026] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1ae4bca-61f0-4729-8888-04ce7b57f0dc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.832150] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 979.832150] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525f2601-bd04-f96b-9bc1-a23736770db2" [ 979.832150] env[61985]: _type = "Task" [ 979.832150] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.841026] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525f2601-bd04-f96b-9bc1-a23736770db2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.907238] env[61985]: DEBUG oslo_vmware.rw_handles [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Completed reading data from the image iterator. {{(pid=61985) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 979.907238] env[61985]: DEBUG oslo_vmware.rw_handles [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52948eac-c8bd-df4c-2c95-6a8663865dba/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 979.907238] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f4d397-15be-46dc-a291-22b03a1e9c3a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.916222] env[61985]: DEBUG oslo_vmware.rw_handles [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52948eac-c8bd-df4c-2c95-6a8663865dba/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 979.916777] env[61985]: DEBUG oslo_vmware.rw_handles [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52948eac-c8bd-df4c-2c95-6a8663865dba/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 979.917196] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-0922a487-9111-4dcb-8c4a-63f72becf5ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.058136] env[61985]: INFO nova.compute.resource_tracker [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating resource usage from migration 1bf48126-5cd8-40dd-bbe0-d7c7e1d54f35 [ 980.095908] env[61985]: DEBUG oslo_vmware.rw_handles [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52948eac-c8bd-df4c-2c95-6a8663865dba/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 980.096740] env[61985]: INFO nova.virt.vmwareapi.images [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Downloaded image file data b4b4e229-a7cf-45de-8912-47f4b206140b [ 980.097409] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a24fcc-d2d8-49c4-82d4-0319664328f3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.123183] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b09ce0e-6cba-4c4b-8aa8-6ed54c143717 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.146162] env[61985]: INFO nova.virt.vmwareapi.images [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] The imported VM was unregistered [ 980.148422] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Caching image {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 980.148730] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating directory with path [datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.149036] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab40726b-f612-4e70-adf8-2a2c91bbbd89 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.180323] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Created directory with path [datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.180562] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda/OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda.vmdk to [datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b/b4b4e229-a7cf-45de-8912-47f4b206140b.vmdk. {{(pid=61985) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 980.181013] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e7b24734-130e-4d8d-a7db-531a70f5fce6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.190362] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 980.190362] env[61985]: value = "task-936235" [ 980.190362] env[61985]: _type = "Task" [ 980.190362] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.200323] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936235, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.258053] env[61985]: DEBUG oslo_concurrency.lockutils [req-4937543a-2f37-4129-a104-d67f8b2930e5 req-d261909c-a6b9-4229-9095-8dad41e25f53 service nova] Releasing lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.276097] env[61985]: DEBUG nova.network.neutron [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance_info_cache with network_info: [{"id": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "address": "fa:16:3e:12:45:ef", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4aa18b7b-89", "ovs_interfaceid": "4aa18b7b-89b5-4cad-af09-1bbc8f255029", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.344363] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525f2601-bd04-f96b-9bc1-a23736770db2, 'name': SearchDatastore_Task, 'duration_secs': 0.016296} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.344685] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.344905] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 980.345159] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.345311] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.345505] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.345778] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-857fb54d-b03d-4331-8515-bc8885678c2b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.354433] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.354613] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 980.355326] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-385cc9db-6c82-4873-bc7e-a2fd1b5aa539 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.361535] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 980.361535] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524c6f52-b3af-c30a-c555-f1b931d4460f" [ 980.361535] env[61985]: _type = "Task" [ 980.361535] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.368431] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524c6f52-b3af-c30a-c555-f1b931d4460f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.391399] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e3e31cc-e1b3-43b7-a6a8-0d34603519d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.400384] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a2969c-6ef6-425c-a740-f31e2041a0a9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.430870] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf613197-950a-4147-b439-dc219e70a559 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.438510] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e246ea-4664-4f4f-b7cf-5069ddbb8d33 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.452534] env[61985]: DEBUG nova.compute.provider_tree [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.638354] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "90ac7d3e-aef3-4a50-963a-606e7c58f446" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.638602] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "90ac7d3e-aef3-4a50-963a-606e7c58f446" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.703105] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936235, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.779097] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-2322ab72-9841-41fb-9d60-2812baabe108" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.871206] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524c6f52-b3af-c30a-c555-f1b931d4460f, 'name': SearchDatastore_Task, 'duration_secs': 0.00818} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.872139] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ddfee5c-c9a6-4000-9e02-ee6a02005005 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.877461] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 980.877461] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c92ee4-7c3f-2b2c-b47b-34bbecb8be90" [ 980.877461] env[61985]: _type = "Task" [ 980.877461] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.887212] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c92ee4-7c3f-2b2c-b47b-34bbecb8be90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.955663] env[61985]: DEBUG nova.scheduler.client.report [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.142046] env[61985]: DEBUG nova.compute.manager [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 981.202577] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936235, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.284677] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 981.284992] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-597b37b9-2f8a-48e9-ada0-2ffd75cabb9d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.293131] env[61985]: DEBUG oslo_vmware.api [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 981.293131] env[61985]: value = "task-936236" [ 981.293131] env[61985]: _type = "Task" [ 981.293131] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.302060] env[61985]: DEBUG oslo_vmware.api [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936236, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.388935] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c92ee4-7c3f-2b2c-b47b-34bbecb8be90, 'name': SearchDatastore_Task, 'duration_secs': 0.068291} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.389355] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.389645] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 5f2aa808-8d1e-471e-9ef7-0f91590ec546/5f2aa808-8d1e-471e-9ef7-0f91590ec546.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 981.389930] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-042426e4-c432-4deb-8297-8fa23ecc3167 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.396676] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 981.396676] env[61985]: value = "task-936237" [ 981.396676] env[61985]: _type = "Task" [ 981.396676] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.405300] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.462086] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.415s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.462086] env[61985]: INFO nova.compute.manager [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Migrating [ 981.468934] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.066s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.470494] env[61985]: INFO nova.compute.claims [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.666433] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.703726] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936235, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.804365] env[61985]: DEBUG oslo_vmware.api [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936236, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.906869] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.984341] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.984837] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.985199] env[61985]: DEBUG nova.network.neutron [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 982.206175] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936235, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.304064] env[61985]: DEBUG oslo_vmware.api [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936236, 'name': PowerOnVM_Task, 'duration_secs': 0.518953} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.304064] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 982.304064] env[61985]: DEBUG nova.compute.manager [None req-3c836826-5b74-4248-86fe-889672c29d86 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.304408] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667958dd-a16d-4f10-a835-5cd8fbbe9a53 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.411250] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.705424] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936235, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.483944} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.708030] env[61985]: INFO nova.virt.vmwareapi.ds_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda/OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda.vmdk to [datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b/b4b4e229-a7cf-45de-8912-47f4b206140b.vmdk. [ 982.708262] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Cleaning up location [datastore2] OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 982.708953] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_4401b671-d426-4eda-9e43-c9bf21941bda {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 982.708953] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40a076be-4280-4e04-aa96-cb93144d5b21 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.715571] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 982.715571] env[61985]: value = "task-936238" [ 982.715571] env[61985]: _type = "Task" [ 982.715571] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.727224] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936238, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.786810] env[61985]: DEBUG nova.network.neutron [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance_info_cache with network_info: [{"id": "262813ed-9baf-4240-b57d-fc4a648b5532", "address": "fa:16:3e:d3:c1:6a", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262813ed-9b", "ovs_interfaceid": "262813ed-9baf-4240-b57d-fc4a648b5532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.845000] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd966071-7874-4575-89b2-e3432c5c54a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.855303] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-706dfb48-bcd7-4650-b32e-8dc4b0e01783 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.886920] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14be6c8-4bcb-4135-a15e-da550e5543a4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.894763] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eef4b34-4aeb-4d4d-a087-505d0f8b1dd9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.913440] env[61985]: DEBUG nova.compute.provider_tree [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.917930] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936237, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.226256] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936238, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217369} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.226535] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.226700] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b/b4b4e229-a7cf-45de-8912-47f4b206140b.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.226951] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b/b4b4e229-a7cf-45de-8912-47f4b206140b.vmdk to [datastore2] 942e0f09-4d34-4aa2-8a60-4c12f899712c/942e0f09-4d34-4aa2-8a60-4c12f899712c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 983.227245] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9104aea1-126a-4086-be54-f4890ed81c40 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.234069] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 983.234069] env[61985]: value = "task-936239" [ 983.234069] env[61985]: _type = "Task" [ 983.234069] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.241562] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936239, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.292432] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.410634] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936237, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.641782} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.410868] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 5f2aa808-8d1e-471e-9ef7-0f91590ec546/5f2aa808-8d1e-471e-9ef7-0f91590ec546.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 983.411096] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 983.411346] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38967647-b68e-4f21-9479-aab010740179 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.417353] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 983.417353] env[61985]: value = "task-936240" [ 983.417353] env[61985]: _type = "Task" [ 983.417353] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.421303] env[61985]: DEBUG nova.scheduler.client.report [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.429684] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936240, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.558726] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "2322ab72-9841-41fb-9d60-2812baabe108" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.559217] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.559570] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "2322ab72-9841-41fb-9d60-2812baabe108-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.559823] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.560136] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.562775] env[61985]: INFO nova.compute.manager [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Terminating instance [ 983.565653] env[61985]: DEBUG nova.compute.manager [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 983.565918] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 983.567208] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b812149-e621-4772-b64c-d98b6c574a9e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.575910] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 983.576267] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d70ae057-b22f-4b20-93a7-5457c8a929ac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.583266] env[61985]: DEBUG oslo_vmware.api [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 983.583266] env[61985]: value = "task-936241" [ 983.583266] env[61985]: _type = "Task" [ 983.583266] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.592785] env[61985]: DEBUG oslo_vmware.api [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936241, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.746897] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936239, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.929580] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.930143] env[61985]: DEBUG nova.compute.manager [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 983.932778] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936240, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.226883} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.933304] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.797s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.934784] env[61985]: INFO nova.compute.claims [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.937425] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 983.938551] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f1ef7c-46ee-4314-ad45-f08197a42d86 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.962206] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 5f2aa808-8d1e-471e-9ef7-0f91590ec546/5f2aa808-8d1e-471e-9ef7-0f91590ec546.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 983.962617] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c7d7010-d62e-4fd4-b70b-cfa2fd212ae5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.984784] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 983.984784] env[61985]: value = "task-936242" [ 983.984784] env[61985]: _type = "Task" [ 983.984784] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.994292] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936242, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.094313] env[61985]: DEBUG oslo_vmware.api [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936241, 'name': PowerOffVM_Task, 'duration_secs': 0.259064} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.094626] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 984.094834] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 984.095146] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a50c7b70-7c11-4fe1-8543-dc1277353a90 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.156179] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 984.156426] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 984.156618] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleting the datastore file [datastore1] 2322ab72-9841-41fb-9d60-2812baabe108 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.156900] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d59a0a8e-3037-4c9b-8f94-3b01f9cfb52c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.163495] env[61985]: DEBUG oslo_vmware.api [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 984.163495] env[61985]: value = "task-936244" [ 984.163495] env[61985]: _type = "Task" [ 984.163495] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.172128] env[61985]: DEBUG oslo_vmware.api [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936244, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.245780] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936239, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.435637] env[61985]: DEBUG nova.compute.utils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.437135] env[61985]: DEBUG nova.compute.manager [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 984.438458] env[61985]: DEBUG nova.network.neutron [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 984.495377] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936242, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.523727] env[61985]: DEBUG nova.policy [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b5abdace901430cb1e431c9a933161e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bdb5f76b9b24a7ba4800a88482847da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.673534] env[61985]: DEBUG oslo_vmware.api [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936244, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.301801} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.673824] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 984.674029] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 984.674369] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 984.674406] env[61985]: INFO nova.compute.manager [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Took 1.11 seconds to destroy the instance on the hypervisor. [ 984.674638] env[61985]: DEBUG oslo.service.loopingcall [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.674842] env[61985]: DEBUG nova.compute.manager [-] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 984.674938] env[61985]: DEBUG nova.network.neutron [-] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 984.745803] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936239, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.810262] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb18298-deb4-4c5e-932a-52aa11aa12da {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.831331] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance '625a3143-d138-4b52-aeb7-9e365c6f1128' progress to 0 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 984.941169] env[61985]: DEBUG nova.compute.manager [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 984.996427] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936242, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.241470] env[61985]: DEBUG nova.network.neutron [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Successfully created port: 4b48329e-7fd7-4641-845a-68423e22587b {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.251142] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936239, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.337479] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 985.337957] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8cc2a75-b1e9-4897-8c8d-25c2b5d30662 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.346327] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 985.346327] env[61985]: value = "task-936245" [ 985.346327] env[61985]: _type = "Task" [ 985.346327] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.355224] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936245, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.442641] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeda7b78-9928-42fa-b1cc-4f0c61c5c6e0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.455129] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a054ad0-12fd-43f6-b6f8-7f68ae64083d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.491620] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2d8725-8cbc-4bc8-be3b-3a5a33011c75 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.507296] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0881755-f020-4f89-b1a6-ac8c90d5ad43 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.511278] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936242, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.522117] env[61985]: DEBUG nova.compute.provider_tree [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.749833] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936239, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.4853} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.749833] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b4b4e229-a7cf-45de-8912-47f4b206140b/b4b4e229-a7cf-45de-8912-47f4b206140b.vmdk to [datastore2] 942e0f09-4d34-4aa2-8a60-4c12f899712c/942e0f09-4d34-4aa2-8a60-4c12f899712c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 985.750683] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ee9e99-06b2-4b58-9508-272c59737a6d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.772773] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 942e0f09-4d34-4aa2-8a60-4c12f899712c/942e0f09-4d34-4aa2-8a60-4c12f899712c.vmdk or device None with type streamOptimized {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.773117] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c5046f7-42be-44a6-8dd2-08eec44a3acb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.793677] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 985.793677] env[61985]: value = "task-936246" [ 985.793677] env[61985]: _type = "Task" [ 985.793677] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.803831] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936246, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.856733] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936245, 'name': PowerOffVM_Task, 'duration_secs': 0.279597} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.857062] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 985.857316] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance '625a3143-d138-4b52-aeb7-9e365c6f1128' progress to 17 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 985.961196] env[61985]: DEBUG nova.compute.manager [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 985.993469] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.993750] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.993954] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.994173] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.994331] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.994488] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.994708] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.994880] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.995166] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.995381] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.995611] env[61985]: DEBUG nova.virt.hardware [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.996496] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58be3b33-1cb8-4788-8094-77ba10dc1634 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.010369] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050ee687-758e-45f0-a059-7632207effd7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.011254] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936242, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.025595] env[61985]: DEBUG nova.scheduler.client.report [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.308692] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936246, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.365353] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.365783] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.366012] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.366287] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.366444] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.366595] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.366804] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.366969] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.367168] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.367341] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.367518] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.372716] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cfc73ae-a1fd-4b11-af0e-bdd9510587c9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.385480] env[61985]: DEBUG nova.compute.manager [req-c3118271-1798-4d1f-ade1-75dc3b778d15 req-cade03b8-6718-4c41-a2e4-320ee154a94d service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Received event network-vif-deleted-4aa18b7b-89b5-4cad-af09-1bbc8f255029 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 986.385480] env[61985]: INFO nova.compute.manager [req-c3118271-1798-4d1f-ade1-75dc3b778d15 req-cade03b8-6718-4c41-a2e4-320ee154a94d service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Neutron deleted interface 4aa18b7b-89b5-4cad-af09-1bbc8f255029; detaching it from the instance and deleting it from the info cache [ 986.385675] env[61985]: DEBUG nova.network.neutron [req-c3118271-1798-4d1f-ade1-75dc3b778d15 req-cade03b8-6718-4c41-a2e4-320ee154a94d service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.392386] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 986.392386] env[61985]: value = "task-936247" [ 986.392386] env[61985]: _type = "Task" [ 986.392386] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.401485] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936247, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.498011] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936242, 'name': ReconfigVM_Task, 'duration_secs': 2.029691} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.498459] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 5f2aa808-8d1e-471e-9ef7-0f91590ec546/5f2aa808-8d1e-471e-9ef7-0f91590ec546.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.499234] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc50fb14-c821-40d1-86ac-0ef2424b381c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.505856] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 986.505856] env[61985]: value = "task-936248" [ 986.505856] env[61985]: _type = "Task" [ 986.505856] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.513573] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936248, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.530995] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.598s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.531573] env[61985]: DEBUG nova.compute.manager [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 986.534233] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.685s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.535709] env[61985]: INFO nova.compute.claims [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.757017] env[61985]: DEBUG nova.network.neutron [-] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.809247] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936246, 'name': ReconfigVM_Task, 'duration_secs': 0.542678} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.809492] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 942e0f09-4d34-4aa2-8a60-4c12f899712c/942e0f09-4d34-4aa2-8a60-4c12f899712c.vmdk or device None with type streamOptimized {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.810211] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dcfbefae-8694-4bbd-a212-2ed2e70de6b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.816676] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 986.816676] env[61985]: value = "task-936249" [ 986.816676] env[61985]: _type = "Task" [ 986.816676] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.826031] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936249, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.888901] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9a4ecef-5938-4896-b9e1-e6b7b4c67e01 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.900919] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e882e09-a485-4b22-895b-7b54873199a3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.914258] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936247, 'name': ReconfigVM_Task, 'duration_secs': 0.459662} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.914911] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance '625a3143-d138-4b52-aeb7-9e365c6f1128' progress to 33 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 986.938368] env[61985]: DEBUG nova.compute.manager [req-c3118271-1798-4d1f-ade1-75dc3b778d15 req-cade03b8-6718-4c41-a2e4-320ee154a94d service nova] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Detach interface failed, port_id=4aa18b7b-89b5-4cad-af09-1bbc8f255029, reason: Instance 2322ab72-9841-41fb-9d60-2812baabe108 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 987.017459] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936248, 'name': Rename_Task, 'duration_secs': 0.210241} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.017700] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 987.017950] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8fbef0d1-fa0d-46bc-b2ea-79cd5a5f929d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.024656] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 987.024656] env[61985]: value = "task-936250" [ 987.024656] env[61985]: _type = "Task" [ 987.024656] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.032229] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936250, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.039691] env[61985]: DEBUG nova.compute.utils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.043028] env[61985]: DEBUG nova.compute.manager [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.043284] env[61985]: DEBUG nova.network.neutron [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 987.092757] env[61985]: DEBUG nova.policy [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faa19f385cd0426d8565e31e286800d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5625013599b44418bd56eb604e14be58', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 987.260952] env[61985]: INFO nova.compute.manager [-] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Took 2.59 seconds to deallocate network for instance. [ 987.326541] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936249, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.422054] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.422054] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.422371] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.422371] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.422533] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.422737] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.422957] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.423151] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.423319] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.423485] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.423720] env[61985]: DEBUG nova.virt.hardware [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.430239] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Reconfiguring VM instance instance-0000004d to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 987.431084] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4f05f49-a347-4e35-9719-7c161b6dc9a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.455684] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 987.455684] env[61985]: value = "task-936251" [ 987.455684] env[61985]: _type = "Task" [ 987.455684] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.465428] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936251, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.535329] env[61985]: DEBUG oslo_vmware.api [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936250, 'name': PowerOnVM_Task, 'duration_secs': 0.443587} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.535677] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 987.535893] env[61985]: INFO nova.compute.manager [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Took 10.99 seconds to spawn the instance on the hypervisor. [ 987.536093] env[61985]: DEBUG nova.compute.manager [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.536849] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c93c43b-3f2e-4e05-9671-9af255cb8842 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.549018] env[61985]: DEBUG nova.compute.manager [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 987.658812] env[61985]: DEBUG nova.network.neutron [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Successfully created port: 8e692ff4-8a38-44ec-ac4b-05e0bda78308 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.771839] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.826827] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936249, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.881031] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c426477f-9ef0-45e4-aa33-b158718005e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.890235] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79329b0-3c8c-49f0-9923-98a20d85f823 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.923557] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77f7da7-134f-426d-8f36-9c7c91fcf0fb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.931603] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f5d3ac-4a63-459a-9b3e-afc552c82210 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.945983] env[61985]: DEBUG nova.compute.provider_tree [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.965295] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936251, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.056667] env[61985]: INFO nova.compute.manager [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Took 39.65 seconds to build instance. [ 988.142518] env[61985]: DEBUG nova.network.neutron [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Successfully updated port: 4b48329e-7fd7-4641-845a-68423e22587b {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.327969] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936249, 'name': Rename_Task, 'duration_secs': 1.213725} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.328276] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 988.328528] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0d3b3ad-c84e-483e-854e-bd5a2968a88b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.334801] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 988.334801] env[61985]: value = "task-936252" [ 988.334801] env[61985]: _type = "Task" [ 988.334801] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.342070] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.448416] env[61985]: DEBUG nova.compute.manager [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Received event network-vif-plugged-4b48329e-7fd7-4641-845a-68423e22587b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 988.448416] env[61985]: DEBUG oslo_concurrency.lockutils [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] Acquiring lock "49707589-4969-4e08-882b-2a2c94bc0d85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.448416] env[61985]: DEBUG oslo_concurrency.lockutils [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] Lock "49707589-4969-4e08-882b-2a2c94bc0d85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.448416] env[61985]: DEBUG oslo_concurrency.lockutils [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] Lock "49707589-4969-4e08-882b-2a2c94bc0d85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.448832] env[61985]: DEBUG nova.compute.manager [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] No waiting events found dispatching network-vif-plugged-4b48329e-7fd7-4641-845a-68423e22587b {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 988.448832] env[61985]: WARNING nova.compute.manager [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Received unexpected event network-vif-plugged-4b48329e-7fd7-4641-845a-68423e22587b for instance with vm_state building and task_state spawning. [ 988.448928] env[61985]: DEBUG nova.compute.manager [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Received event network-changed-4b48329e-7fd7-4641-845a-68423e22587b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 988.449162] env[61985]: DEBUG nova.compute.manager [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Refreshing instance network info cache due to event network-changed-4b48329e-7fd7-4641-845a-68423e22587b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 988.449477] env[61985]: DEBUG oslo_concurrency.lockutils [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] Acquiring lock "refresh_cache-49707589-4969-4e08-882b-2a2c94bc0d85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.449677] env[61985]: DEBUG oslo_concurrency.lockutils [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] Acquired lock "refresh_cache-49707589-4969-4e08-882b-2a2c94bc0d85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.449882] env[61985]: DEBUG nova.network.neutron [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Refreshing network info cache for port 4b48329e-7fd7-4641-845a-68423e22587b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 988.451783] env[61985]: DEBUG nova.scheduler.client.report [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.467995] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936251, 'name': ReconfigVM_Task, 'duration_secs': 0.751798} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.467995] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Reconfigured VM instance instance-0000004d to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 988.468400] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032b20e5-68fb-4c4c-a87e-3bafa84ec57e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.490606] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 625a3143-d138-4b52-aeb7-9e365c6f1128/625a3143-d138-4b52-aeb7-9e365c6f1128.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.491314] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46dbb538-abea-4d28-8be2-b6b0bf7bf3f7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.509440] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 988.509440] env[61985]: value = "task-936253" [ 988.509440] env[61985]: _type = "Task" [ 988.509440] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.519280] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936253, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.557805] env[61985]: DEBUG nova.compute.manager [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 988.560785] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a4683d7-650e-40c1-ba94-238f718eaa0d tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.164s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.584112] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.584414] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.584589] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.584814] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.584979] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.585162] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.585378] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.585543] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.585717] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.585886] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.586082] env[61985]: DEBUG nova.virt.hardware [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.586917] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29eeb2e-7486-4e69-a926-46fd1623b200 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.594874] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4083d99c-ec0c-4462-9a3b-70689e3d26a8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.645752] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "refresh_cache-49707589-4969-4e08-882b-2a2c94bc0d85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.844935] env[61985]: DEBUG oslo_vmware.api [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936252, 'name': PowerOnVM_Task, 'duration_secs': 0.408684} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.845268] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 988.845481] env[61985]: INFO nova.compute.manager [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Took 14.81 seconds to spawn the instance on the hypervisor. [ 988.845705] env[61985]: DEBUG nova.compute.manager [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.846481] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b539450-7c3d-4367-a12f-800e975c9de7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.958433] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.959087] env[61985]: DEBUG nova.compute.manager [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 988.961912] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 24.150s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.019979] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936253, 'name': ReconfigVM_Task, 'duration_secs': 0.284542} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.020785] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 625a3143-d138-4b52-aeb7-9e365c6f1128/625a3143-d138-4b52-aeb7-9e365c6f1128.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.021088] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance '625a3143-d138-4b52-aeb7-9e365c6f1128' progress to 50 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 989.044917] env[61985]: DEBUG nova.network.neutron [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 989.259851] env[61985]: DEBUG nova.network.neutron [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.369583] env[61985]: INFO nova.compute.manager [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Took 43.19 seconds to build instance. [ 989.467103] env[61985]: DEBUG nova.compute.utils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 989.469268] env[61985]: DEBUG nova.objects.instance [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lazy-loading 'migration_context' on Instance uuid a77ca23f-b2c0-4822-8e48-3e47e0dadb27 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.470540] env[61985]: DEBUG nova.compute.manager [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 989.470708] env[61985]: DEBUG nova.network.neutron [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 989.532412] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a76c21-7db2-45fb-9658-2a89a6a50ee4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.556678] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b60703-908f-4109-9625-9980c6002997 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.574694] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance '625a3143-d138-4b52-aeb7-9e365c6f1128' progress to 67 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 989.763219] env[61985]: DEBUG oslo_concurrency.lockutils [req-c1515466-e652-4310-ae63-83b6dbc1a44b req-80833277-62ef-4e3b-858d-271f5d465745 service nova] Releasing lock "refresh_cache-49707589-4969-4e08-882b-2a2c94bc0d85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.763600] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "refresh_cache-49707589-4969-4e08-882b-2a2c94bc0d85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.763761] env[61985]: DEBUG nova.network.neutron [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 989.804614] env[61985]: DEBUG nova.compute.manager [req-c57023ab-50be-4ad9-a1de-0f8cb5d829bd req-da6ada4f-3128-47ad-8fe2-51f88f778248 service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Received event network-vif-plugged-8e692ff4-8a38-44ec-ac4b-05e0bda78308 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 989.804853] env[61985]: DEBUG oslo_concurrency.lockutils [req-c57023ab-50be-4ad9-a1de-0f8cb5d829bd req-da6ada4f-3128-47ad-8fe2-51f88f778248 service nova] Acquiring lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.805104] env[61985]: DEBUG oslo_concurrency.lockutils [req-c57023ab-50be-4ad9-a1de-0f8cb5d829bd req-da6ada4f-3128-47ad-8fe2-51f88f778248 service nova] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.805291] env[61985]: DEBUG oslo_concurrency.lockutils [req-c57023ab-50be-4ad9-a1de-0f8cb5d829bd req-da6ada4f-3128-47ad-8fe2-51f88f778248 service nova] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.805469] env[61985]: DEBUG nova.compute.manager [req-c57023ab-50be-4ad9-a1de-0f8cb5d829bd req-da6ada4f-3128-47ad-8fe2-51f88f778248 service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] No waiting events found dispatching network-vif-plugged-8e692ff4-8a38-44ec-ac4b-05e0bda78308 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.805761] env[61985]: WARNING nova.compute.manager [req-c57023ab-50be-4ad9-a1de-0f8cb5d829bd req-da6ada4f-3128-47ad-8fe2-51f88f778248 service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Received unexpected event network-vif-plugged-8e692ff4-8a38-44ec-ac4b-05e0bda78308 for instance with vm_state building and task_state spawning. [ 989.816195] env[61985]: DEBUG nova.policy [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e23fdb33638143409613d9d4f26f199c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe531ec9f60548adb457351e19f8d51b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 989.871806] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9f2415db-f240-40b5-83ad-92ee1b2b8b5e tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "942e0f09-4d34-4aa2-8a60-4c12f899712c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.704s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.942142] env[61985]: DEBUG nova.network.neutron [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Successfully updated port: 8e692ff4-8a38-44ec-ac4b-05e0bda78308 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.970267] env[61985]: DEBUG nova.compute.manager [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 990.216763] env[61985]: DEBUG nova.network.neutron [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Port 262813ed-9baf-4240-b57d-fc4a648b5532 binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 990.327085] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf5c55d-d0f3-4908-99dd-992544fb9be2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.335360] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3614b9ba-c53b-467c-a3cd-24124cb36f0b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.368640] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256fa119-aeb0-4816-af9e-3fb2ebfdf2b7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.373800] env[61985]: DEBUG nova.network.neutron [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 990.379353] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e1aeec-9bbb-4ce8-af1d-0bb1445c8773 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.394404] env[61985]: DEBUG nova.compute.provider_tree [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.450295] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.450483] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.450642] env[61985]: DEBUG nova.network.neutron [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 990.690391] env[61985]: DEBUG nova.network.neutron [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Successfully created port: 9f39fdcd-6822-4b1e-b8c8-63d8c3878936 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 990.754247] env[61985]: DEBUG nova.network.neutron [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Updating instance_info_cache with network_info: [{"id": "4b48329e-7fd7-4641-845a-68423e22587b", "address": "fa:16:3e:e5:10:72", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b48329e-7f", "ovs_interfaceid": "4b48329e-7fd7-4641-845a-68423e22587b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.852018] env[61985]: DEBUG nova.compute.manager [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Received event network-changed-8e692ff4-8a38-44ec-ac4b-05e0bda78308 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 990.852339] env[61985]: DEBUG nova.compute.manager [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Refreshing instance network info cache due to event network-changed-8e692ff4-8a38-44ec-ac4b-05e0bda78308. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 990.852530] env[61985]: DEBUG oslo_concurrency.lockutils [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] Acquiring lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.897893] env[61985]: DEBUG nova.scheduler.client.report [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.984520] env[61985]: DEBUG nova.compute.manager [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 991.005301] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.005565] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.005727] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.005915] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.006107] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.006276] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.006607] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.006799] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.006976] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.007239] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.007469] env[61985]: DEBUG nova.virt.hardware [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.011506] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b4caee-c21c-4ea3-a625-edde83b6937d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.020707] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dcd67de-5b70-4a23-9992-2fde3c18fe1f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.042931] env[61985]: DEBUG nova.network.neutron [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 991.095997] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "942e0f09-4d34-4aa2-8a60-4c12f899712c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.096310] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "942e0f09-4d34-4aa2-8a60-4c12f899712c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.096527] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "942e0f09-4d34-4aa2-8a60-4c12f899712c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.096716] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "942e0f09-4d34-4aa2-8a60-4c12f899712c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.096891] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "942e0f09-4d34-4aa2-8a60-4c12f899712c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.099349] env[61985]: INFO nova.compute.manager [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Terminating instance [ 991.101410] env[61985]: DEBUG nova.compute.manager [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 991.101485] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 991.102705] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033e61dd-fde2-4f48-9196-3307e39bb62d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.114306] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 991.114538] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a293fe0-0539-4e9e-a541-0eb4855c4f30 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.121168] env[61985]: DEBUG oslo_vmware.api [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 991.121168] env[61985]: value = "task-936254" [ 991.121168] env[61985]: _type = "Task" [ 991.121168] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.128772] env[61985]: DEBUG oslo_vmware.api [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936254, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.246021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "625a3143-d138-4b52-aeb7-9e365c6f1128-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.246021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.246021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.257861] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "refresh_cache-49707589-4969-4e08-882b-2a2c94bc0d85" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.258200] env[61985]: DEBUG nova.compute.manager [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Instance network_info: |[{"id": "4b48329e-7fd7-4641-845a-68423e22587b", "address": "fa:16:3e:e5:10:72", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b48329e-7f", "ovs_interfaceid": "4b48329e-7fd7-4641-845a-68423e22587b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 991.258645] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e5:10:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b48329e-7fd7-4641-845a-68423e22587b', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.267692] env[61985]: DEBUG oslo.service.loopingcall [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.269645] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 991.269938] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f329fcb4-9eb0-4676-be93-b7d8a759a93e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.293439] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.293439] env[61985]: value = "task-936255" [ 991.293439] env[61985]: _type = "Task" [ 991.293439] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.302199] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936255, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.460700] env[61985]: DEBUG nova.network.neutron [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Updating instance_info_cache with network_info: [{"id": "8e692ff4-8a38-44ec-ac4b-05e0bda78308", "address": "fa:16:3e:80:98:79", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e692ff4-8a", "ovs_interfaceid": "8e692ff4-8a38-44ec-ac4b-05e0bda78308", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.632783] env[61985]: DEBUG oslo_vmware.api [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936254, 'name': PowerOffVM_Task, 'duration_secs': 0.337789} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.635224] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.635666] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 991.635905] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e0d935a-2a61-43e6-ad9e-29a11abc428f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.760425] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 991.760699] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 991.760842] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleting the datastore file [datastore2] 942e0f09-4d34-4aa2-8a60-4c12f899712c {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.761372] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44b1e697-e675-4fb2-ae7a-948fc2e9a4c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.769380] env[61985]: DEBUG oslo_vmware.api [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 991.769380] env[61985]: value = "task-936257" [ 991.769380] env[61985]: _type = "Task" [ 991.769380] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.777286] env[61985]: DEBUG oslo_vmware.api [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936257, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.802399] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936255, 'name': CreateVM_Task, 'duration_secs': 0.484294} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.802565] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 991.803234] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.803408] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.803760] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 991.803974] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9856a92-ed83-47db-a666-71a03d06bed4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.808227] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 991.808227] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52691703-aabf-1203-cad7-865ad3ccce99" [ 991.808227] env[61985]: _type = "Task" [ 991.808227] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.815443] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52691703-aabf-1203-cad7-865ad3ccce99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.909707] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.948s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.915275] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 24.848s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.915459] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.915649] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 991.915973] env[61985]: DEBUG oslo_concurrency.lockutils [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.191s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.916213] env[61985]: DEBUG nova.objects.instance [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lazy-loading 'resources' on Instance uuid 57591f6f-ed7d-46f3-a9c1-217016c42b1c {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.919963] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b406b1-3b6f-4a0d-9c09-492253774ec8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.937730] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c17ebcc8-0dc5-4ec3-a862-e38b50babbd0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.955284] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6ed473-8dfe-4275-8cda-587cf0914c82 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.961994] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.962316] env[61985]: DEBUG nova.compute.manager [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Instance network_info: |[{"id": "8e692ff4-8a38-44ec-ac4b-05e0bda78308", "address": "fa:16:3e:80:98:79", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e692ff4-8a", "ovs_interfaceid": "8e692ff4-8a38-44ec-ac4b-05e0bda78308", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 991.962667] env[61985]: DEBUG oslo_concurrency.lockutils [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] Acquired lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.962855] env[61985]: DEBUG nova.network.neutron [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Refreshing network info cache for port 8e692ff4-8a38-44ec-ac4b-05e0bda78308 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 991.963939] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:98:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7908211b-df93-467b-87a8-3c3d29b03de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e692ff4-8a38-44ec-ac4b-05e0bda78308', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.971260] env[61985]: DEBUG oslo.service.loopingcall [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.972227] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f5ee3b-8f85-4278-b96b-bbb02c33d5af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.980366] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 991.980989] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c75e758-d4c4-4172-8228-d7abb6bdf7ba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.022021] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179627MB free_disk=43GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 992.022199] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.027929] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.027929] env[61985]: value = "task-936258" [ 992.027929] env[61985]: _type = "Task" [ 992.027929] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.035689] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936258, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.065462] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "8db82979-e527-4c63-8ce8-fe1456f7708c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.065739] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "8db82979-e527-4c63-8ce8-fe1456f7708c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.213234] env[61985]: DEBUG nova.network.neutron [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Updated VIF entry in instance network info cache for port 8e692ff4-8a38-44ec-ac4b-05e0bda78308. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 992.213630] env[61985]: DEBUG nova.network.neutron [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Updating instance_info_cache with network_info: [{"id": "8e692ff4-8a38-44ec-ac4b-05e0bda78308", "address": "fa:16:3e:80:98:79", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e692ff4-8a", "ovs_interfaceid": "8e692ff4-8a38-44ec-ac4b-05e0bda78308", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.279352] env[61985]: DEBUG oslo_vmware.api [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936257, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140257} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.281944] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.282189] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 992.282388] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 992.282576] env[61985]: INFO nova.compute.manager [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Took 1.18 seconds to destroy the instance on the hypervisor. [ 992.282917] env[61985]: DEBUG oslo.service.loopingcall [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.283383] env[61985]: DEBUG nova.compute.manager [-] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.283491] env[61985]: DEBUG nova.network.neutron [-] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 992.293413] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42bf00b1-21f1-481a-8165-dc902e8afc5e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.300917] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6914ee09-a006-4928-80d2-e01517b61663 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.305152] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.305372] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.305680] env[61985]: DEBUG nova.network.neutron [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 992.338311] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0956ae83-b4c1-4234-8a8e-0c4aaf8ee96d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.346481] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52691703-aabf-1203-cad7-865ad3ccce99, 'name': SearchDatastore_Task, 'duration_secs': 0.007982} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.347350] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.347618] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 992.348145] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.348145] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.348276] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.348684] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc6c3fc1-41c1-4d08-addc-de43d66b4a58 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.353446] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd955d23-c34e-4f20-9f63-0967225555e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.358458] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 992.358811] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 992.359604] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7e27ba6-f32a-4da3-bee5-995dbfdf081f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.372609] env[61985]: DEBUG nova.compute.provider_tree [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.375026] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 992.375026] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a35178-fd75-a30b-46fa-e8923d5e7a01" [ 992.375026] env[61985]: _type = "Task" [ 992.375026] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.382057] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a35178-fd75-a30b-46fa-e8923d5e7a01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.543196] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936258, 'name': CreateVM_Task, 'duration_secs': 0.307915} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.543378] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 992.544101] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.544264] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.544588] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 992.544841] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ca0d45f-c996-4ced-84f7-d67ae674a845 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.550091] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 992.550091] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d2943c-135c-f2b5-31ea-b14790353f7b" [ 992.550091] env[61985]: _type = "Task" [ 992.550091] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.558565] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d2943c-135c-f2b5-31ea-b14790353f7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.564912] env[61985]: DEBUG nova.compute.manager [req-a8a9f900-28d2-4377-9b7f-78e3eab89491 req-b73f194e-f120-4c35-a7c9-262ada0fd2d4 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Received event network-vif-deleted-c7cba622-8372-4f2f-994a-8c92dab641bb {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 992.566169] env[61985]: INFO nova.compute.manager [req-a8a9f900-28d2-4377-9b7f-78e3eab89491 req-b73f194e-f120-4c35-a7c9-262ada0fd2d4 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Neutron deleted interface c7cba622-8372-4f2f-994a-8c92dab641bb; detaching it from the instance and deleting it from the info cache [ 992.566169] env[61985]: DEBUG nova.network.neutron [req-a8a9f900-28d2-4377-9b7f-78e3eab89491 req-b73f194e-f120-4c35-a7c9-262ada0fd2d4 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.568086] env[61985]: DEBUG nova.compute.manager [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 992.718603] env[61985]: DEBUG oslo_concurrency.lockutils [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] Releasing lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.718876] env[61985]: DEBUG nova.compute.manager [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received event network-changed-7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 992.719060] env[61985]: DEBUG nova.compute.manager [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing instance network info cache due to event network-changed-7075efce-177a-4127-9c2f-7e3a9d1e92ad. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 992.719287] env[61985]: DEBUG oslo_concurrency.lockutils [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] Acquiring lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.719439] env[61985]: DEBUG oslo_concurrency.lockutils [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] Acquired lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.719606] env[61985]: DEBUG nova.network.neutron [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing network info cache for port 7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 992.765492] env[61985]: DEBUG nova.network.neutron [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Successfully updated port: 9f39fdcd-6822-4b1e-b8c8-63d8c3878936 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 992.877098] env[61985]: DEBUG nova.scheduler.client.report [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.887682] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a35178-fd75-a30b-46fa-e8923d5e7a01, 'name': SearchDatastore_Task, 'duration_secs': 0.008072} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.888768] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d98f698-8e04-4220-96ca-011040da9177 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.895481] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 992.895481] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b82cb9-168d-08f7-006f-0796801215aa" [ 992.895481] env[61985]: _type = "Task" [ 992.895481] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.904840] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b82cb9-168d-08f7-006f-0796801215aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.957018] env[61985]: DEBUG nova.compute.manager [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Received event network-vif-plugged-9f39fdcd-6822-4b1e-b8c8-63d8c3878936 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 992.957132] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] Acquiring lock "b6eb50ff-e685-4e8a-92f5-c2661136c361-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.957385] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] Lock "b6eb50ff-e685-4e8a-92f5-c2661136c361-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.957512] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] Lock "b6eb50ff-e685-4e8a-92f5-c2661136c361-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.957717] env[61985]: DEBUG nova.compute.manager [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] No waiting events found dispatching network-vif-plugged-9f39fdcd-6822-4b1e-b8c8-63d8c3878936 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.957892] env[61985]: WARNING nova.compute.manager [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Received unexpected event network-vif-plugged-9f39fdcd-6822-4b1e-b8c8-63d8c3878936 for instance with vm_state building and task_state spawning. [ 992.958070] env[61985]: DEBUG nova.compute.manager [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Received event network-changed-9f39fdcd-6822-4b1e-b8c8-63d8c3878936 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 992.958232] env[61985]: DEBUG nova.compute.manager [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Refreshing instance network info cache due to event network-changed-9f39fdcd-6822-4b1e-b8c8-63d8c3878936. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 992.958419] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] Acquiring lock "refresh_cache-b6eb50ff-e685-4e8a-92f5-c2661136c361" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.958559] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] Acquired lock "refresh_cache-b6eb50ff-e685-4e8a-92f5-c2661136c361" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.958801] env[61985]: DEBUG nova.network.neutron [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Refreshing network info cache for port 9f39fdcd-6822-4b1e-b8c8-63d8c3878936 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 993.036917] env[61985]: DEBUG nova.network.neutron [-] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.064887] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d2943c-135c-f2b5-31ea-b14790353f7b, 'name': SearchDatastore_Task, 'duration_secs': 0.010132} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.066068] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.066323] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.066539] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.067799] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6555bf2e-a091-4a83-bda0-58c9b89a6609 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.080146] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb4a288-d9a6-4ed8-8390-ab3bac90fe24 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.106674] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.121290] env[61985]: DEBUG nova.compute.manager [req-a8a9f900-28d2-4377-9b7f-78e3eab89491 req-b73f194e-f120-4c35-a7c9-262ada0fd2d4 service nova] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Detach interface failed, port_id=c7cba622-8372-4f2f-994a-8c92dab641bb, reason: Instance 942e0f09-4d34-4aa2-8a60-4c12f899712c could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 993.166055] env[61985]: DEBUG nova.network.neutron [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance_info_cache with network_info: [{"id": "262813ed-9baf-4240-b57d-fc4a648b5532", "address": "fa:16:3e:d3:c1:6a", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262813ed-9b", "ovs_interfaceid": "262813ed-9baf-4240-b57d-fc4a648b5532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.267621] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "refresh_cache-b6eb50ff-e685-4e8a-92f5-c2661136c361" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.382133] env[61985]: DEBUG oslo_concurrency.lockutils [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.466s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.384363] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.750s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.385850] env[61985]: INFO nova.compute.claims [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.409747] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b82cb9-168d-08f7-006f-0796801215aa, 'name': SearchDatastore_Task, 'duration_secs': 0.023332} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.410043] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.410368] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 49707589-4969-4e08-882b-2a2c94bc0d85/49707589-4969-4e08-882b-2a2c94bc0d85.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 993.411348] env[61985]: INFO nova.scheduler.client.report [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleted allocations for instance 57591f6f-ed7d-46f3-a9c1-217016c42b1c [ 993.412491] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.412601] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.412824] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f255dcb6-a4dc-43db-8a1d-79055048e46b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.417093] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8e2a16e-9a6e-4f15-9b1a-781593e3cb20 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.425658] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 993.425658] env[61985]: value = "task-936259" [ 993.425658] env[61985]: _type = "Task" [ 993.425658] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.429923] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.430143] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 993.431231] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de3cffb0-a06c-4279-9e9b-83064bd669ca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.437398] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936259, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.444279] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 993.444279] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5229bc9a-bfa3-a497-2a2e-f8524aee5cfc" [ 993.444279] env[61985]: _type = "Task" [ 993.444279] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.453745] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5229bc9a-bfa3-a497-2a2e-f8524aee5cfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.457643] env[61985]: INFO nova.compute.manager [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Swapping old allocation on dict_keys(['aed7e5b3-c662-4538-8447-c4f67b460215']) held by migration 9da5e5c0-4f88-4a0e-98ee-ea219e2fc85d for instance [ 993.475484] env[61985]: DEBUG nova.network.neutron [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updated VIF entry in instance network info cache for port 7075efce-177a-4127-9c2f-7e3a9d1e92ad. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 993.476010] env[61985]: DEBUG nova.network.neutron [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [{"id": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "address": "fa:16:3e:4b:3d:56", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7075efce-17", "ovs_interfaceid": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.482143] env[61985]: DEBUG nova.scheduler.client.report [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Overwriting current allocation {'allocations': {'aed7e5b3-c662-4538-8447-c4f67b460215': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 114}}, 'project_id': '643e8788ab5746cb9d3259031c275261', 'user_id': '248945e574d34c10a4f2387688673613', 'consumer_generation': 1} on consumer a77ca23f-b2c0-4822-8e48-3e47e0dadb27 {{(pid=61985) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 993.493355] env[61985]: DEBUG nova.network.neutron [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 993.539866] env[61985]: INFO nova.compute.manager [-] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Took 1.26 seconds to deallocate network for instance. [ 993.563319] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.563513] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquired lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.563691] env[61985]: DEBUG nova.network.neutron [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 993.569203] env[61985]: DEBUG nova.network.neutron [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.669186] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.923468] env[61985]: DEBUG oslo_concurrency.lockutils [None req-32364615-e8ab-4928-ac57-02784f0c7b2e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "57591f6f-ed7d-46f3-a9c1-217016c42b1c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.213s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.953123] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936259, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.957318] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5229bc9a-bfa3-a497-2a2e-f8524aee5cfc, 'name': SearchDatastore_Task, 'duration_secs': 0.012214} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.958203] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c19512c-b227-4035-92d2-b00c60cccc9e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.964576] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 993.964576] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523bf3c3-e5da-c518-a6cf-6ddf4de94d6b" [ 993.964576] env[61985]: _type = "Task" [ 993.964576] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.974050] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523bf3c3-e5da-c518-a6cf-6ddf4de94d6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.978680] env[61985]: DEBUG oslo_concurrency.lockutils [req-22d33f91-cac8-4156-8361-ca53bb1a2ae2 req-96c0177e-f60d-468d-a527-4f022de861a8 service nova] Releasing lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.046156] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.074979] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5e59492-5f98-43e8-80d0-9272e2a33a38 req-c0003b0e-bdec-4ef7-9c8a-d28f91680ac8 service nova] Releasing lock "refresh_cache-b6eb50ff-e685-4e8a-92f5-c2661136c361" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.075403] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "refresh_cache-b6eb50ff-e685-4e8a-92f5-c2661136c361" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.075574] env[61985]: DEBUG nova.network.neutron [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 994.193784] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a007c4-40cb-4068-8e05-0250f8e88662 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.215985] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d432e041-e5d6-4190-8d5b-d4cfe39591b2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.224712] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance '625a3143-d138-4b52-aeb7-9e365c6f1128' progress to 83 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 994.266187] env[61985]: DEBUG nova.network.neutron [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance_info_cache with network_info: [{"id": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "address": "fa:16:3e:b5:b4:b7", "network": {"id": "b0948f09-f4b2-43ba-851c-6d49beca2e30", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.147", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "5c359de6550041d28450fc8d15586c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d016adf-58", "ovs_interfaceid": "1d016adf-58f0-4fc2-a9d5-4e7232356db4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.439217] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936259, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555587} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.441636] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 49707589-4969-4e08-882b-2a2c94bc0d85/49707589-4969-4e08-882b-2a2c94bc0d85.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 994.441870] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 994.442412] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8db0bdd5-1830-4ddb-ab92-96ff9eff747b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.449213] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 994.449213] env[61985]: value = "task-936260" [ 994.449213] env[61985]: _type = "Task" [ 994.449213] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.462479] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936260, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.477133] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523bf3c3-e5da-c518-a6cf-6ddf4de94d6b, 'name': SearchDatastore_Task, 'duration_secs': 0.00908} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.479692] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.479972] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] afa46da0-3bd1-47ea-bdf0-28eed87fd627/afa46da0-3bd1-47ea-bdf0-28eed87fd627.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 994.480625] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c531bce-1179-4b4d-aba3-af241dc80e03 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.488614] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 994.488614] env[61985]: value = "task-936261" [ 994.488614] env[61985]: _type = "Task" [ 994.488614] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.501609] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936261, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.604992] env[61985]: DEBUG nova.network.neutron [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 994.724385] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd623d6b-40d7-4f5c-bcf0-308da4a72cff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.731749] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 994.734185] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb6cfc4a-6f1c-40b8-8b33-761b3d270cce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.737083] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be6a302f-382d-41dd-8ee8-7dea457b66d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.771543] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3cde953-3cdb-4b3d-8906-b8e0e0fc1d3e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.774637] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Releasing lock "refresh_cache-a77ca23f-b2c0-4822-8e48-3e47e0dadb27" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.775127] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 994.775471] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 994.775471] env[61985]: value = "task-936262" [ 994.775471] env[61985]: _type = "Task" [ 994.775471] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.775734] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68341e27-8d69-4a87-97f1-1ca275e15549 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.785416] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a02c3ecc-d4c3-4899-8db0-68367b6be594 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.795353] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936262, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.795678] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 994.795678] env[61985]: value = "task-936263" [ 994.795678] env[61985]: _type = "Task" [ 994.795678] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.800790] env[61985]: DEBUG nova.network.neutron [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Updating instance_info_cache with network_info: [{"id": "9f39fdcd-6822-4b1e-b8c8-63d8c3878936", "address": "fa:16:3e:78:01:df", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f39fdcd-68", "ovs_interfaceid": "9f39fdcd-6822-4b1e-b8c8-63d8c3878936", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.811122] env[61985]: DEBUG nova.compute.provider_tree [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.815750] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.963250] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "6dd8218b-c14d-40c9-87df-097fab06c669" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.963696] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.964999] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936260, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067594} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.965601] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.966531] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de1a4bf-8983-4f74-a67f-340c945a2e13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.996858] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 49707589-4969-4e08-882b-2a2c94bc0d85/49707589-4969-4e08-882b-2a2c94bc0d85.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.998067] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d6673a9-b82a-47f2-a368-7ee2ae2ba19c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.023022] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936261, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.025127] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 995.025127] env[61985]: value = "task-936264" [ 995.025127] env[61985]: _type = "Task" [ 995.025127] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.035375] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936264, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.291548] env[61985]: DEBUG oslo_vmware.api [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936262, 'name': PowerOnVM_Task, 'duration_secs': 0.450433} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.295223] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 995.295223] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dd018bdd-a95f-488e-9f8c-47bb29735b52 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance '625a3143-d138-4b52-aeb7-9e365c6f1128' progress to 100 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 995.307206] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936263, 'name': PowerOffVM_Task, 'duration_secs': 0.225764} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.307516] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 995.308316] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:29:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='d8c8cc63-01d7-4c63-90a9-d468b4fc7cea',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1410428606',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.308533] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.308742] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.308974] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.309140] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.309291] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.309504] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.309720] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.309935] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.310188] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.310297] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.316460] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "refresh_cache-b6eb50ff-e685-4e8a-92f5-c2661136c361" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.316760] env[61985]: DEBUG nova.compute.manager [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Instance network_info: |[{"id": "9f39fdcd-6822-4b1e-b8c8-63d8c3878936", "address": "fa:16:3e:78:01:df", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f39fdcd-68", "ovs_interfaceid": "9f39fdcd-6822-4b1e-b8c8-63d8c3878936", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 995.317070] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e0630e1-8b1f-421a-ba00-384c84cc1fc3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.328757] env[61985]: DEBUG nova.scheduler.client.report [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.332240] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:01:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55520f67-d092-4eb7-940f-d7cceaa1ca1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f39fdcd-6822-4b1e-b8c8-63d8c3878936', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.340620] env[61985]: DEBUG oslo.service.loopingcall [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.341210] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 995.342415] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a584dc3-b9e2-43dd-80b9-044dd21d7dd7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.359674] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 995.359674] env[61985]: value = "task-936265" [ 995.359674] env[61985]: _type = "Task" [ 995.359674] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.365017] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.365017] env[61985]: value = "task-936266" [ 995.365017] env[61985]: _type = "Task" [ 995.365017] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.371835] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936265, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.379699] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936266, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.469020] env[61985]: INFO nova.compute.manager [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Detaching volume 7edc010c-38ef-42c0-af8a-550bed9ea645 [ 995.503247] env[61985]: INFO nova.virt.block_device [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Attempting to driver detach volume 7edc010c-38ef-42c0-af8a-550bed9ea645 from mountpoint /dev/sdb [ 995.503527] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 995.503761] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211422', 'volume_id': '7edc010c-38ef-42c0-af8a-550bed9ea645', 'name': 'volume-7edc010c-38ef-42c0-af8a-550bed9ea645', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dd8218b-c14d-40c9-87df-097fab06c669', 'attached_at': '', 'detached_at': '', 'volume_id': '7edc010c-38ef-42c0-af8a-550bed9ea645', 'serial': '7edc010c-38ef-42c0-af8a-550bed9ea645'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 995.504873] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485a447f-bb9b-4e8d-afb9-e34d1c5c2ffb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.514419] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936261, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.705704} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.532948] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] afa46da0-3bd1-47ea-bdf0-28eed87fd627/afa46da0-3bd1-47ea-bdf0-28eed87fd627.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 995.533286] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 995.533723] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-28b7f7f0-e9a5-4a5f-a589-251b876929c2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.539663] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d1a93b-0b7e-4087-ac93-55c16a8c6f13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.550791] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936264, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.553704] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 995.553704] env[61985]: value = "task-936267" [ 995.553704] env[61985]: _type = "Task" [ 995.553704] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.554644] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0f0f9a-ea38-4566-bfaf-40a35fd20902 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.567503] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936267, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.584936] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edbe68b-abf9-4c41-935f-564d5982f6cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.604591] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] The volume has not been displaced from its original location: [datastore2] volume-7edc010c-38ef-42c0-af8a-550bed9ea645/volume-7edc010c-38ef-42c0-af8a-550bed9ea645.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 995.610236] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Reconfiguring VM instance instance-00000043 to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 995.610626] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d963f66-418a-4923-a937-47acc9e2d97e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.629950] env[61985]: DEBUG oslo_vmware.api [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 995.629950] env[61985]: value = "task-936268" [ 995.629950] env[61985]: _type = "Task" [ 995.629950] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.638935] env[61985]: DEBUG oslo_vmware.api [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936268, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.845081] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.845081] env[61985]: DEBUG nova.compute.manager [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 995.845799] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.488s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.850732] env[61985]: DEBUG nova.objects.instance [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lazy-loading 'resources' on Instance uuid 4da2d122-aa36-4cad-9521-498b53b9f2f6 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.888025] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936265, 'name': ReconfigVM_Task, 'duration_secs': 0.261584} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.888610] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936266, 'name': CreateVM_Task, 'duration_secs': 0.462242} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.889531] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e2cc67-a38b-49c0-a300-52f4aae4cfe0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.892210] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 995.893313] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.893616] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.894061] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.894779] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2caf0bb8-bd09-42cb-bc4f-93096db2d859 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.911683] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:29:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='d8c8cc63-01d7-4c63-90a9-d468b4fc7cea',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1410428606',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.912240] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.912528] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.912826] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.913108] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.913397] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.913750] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.914072] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.914371] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.914686] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.915024] env[61985]: DEBUG nova.virt.hardware [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.917526] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80329e46-e898-4a5e-894f-3aeee4a4fa30 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.920325] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 995.920325] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5216b8e4-09c4-4788-6f7d-965f629c42a7" [ 995.920325] env[61985]: _type = "Task" [ 995.920325] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.927507] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 995.927507] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52713cf3-975a-14a2-9591-cd1b3665ddee" [ 995.927507] env[61985]: _type = "Task" [ 995.927507] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.952449] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5216b8e4-09c4-4788-6f7d-965f629c42a7, 'name': SearchDatastore_Task, 'duration_secs': 0.011408} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.956485] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.956974] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.957378] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.957710] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.958035] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.958433] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52713cf3-975a-14a2-9591-cd1b3665ddee, 'name': SearchDatastore_Task, 'duration_secs': 0.024485} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.958789] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5cef4f55-6966-4987-9d12-13dc02d2ced9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.965827] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Reconfiguring VM instance instance-00000047 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 995.966469] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a177543-b479-45bc-8457-23744b1c7c3f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.988050] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 995.988050] env[61985]: value = "task-936269" [ 995.988050] env[61985]: _type = "Task" [ 995.988050] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.989595] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.989777] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 995.993564] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a031899b-ad33-439e-bf53-447ce289bf03 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.999601] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 995.999601] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529111f8-07e4-8592-b254-8cee973b9240" [ 995.999601] env[61985]: _type = "Task" [ 995.999601] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.003088] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936269, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.012450] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529111f8-07e4-8592-b254-8cee973b9240, 'name': SearchDatastore_Task} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.013255] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1486b7d3-7ebd-4304-9fce-04e0c3331819 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.019508] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 996.019508] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520e3842-55cb-fd6b-b85e-183c1a66a24b" [ 996.019508] env[61985]: _type = "Task" [ 996.019508] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.028194] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520e3842-55cb-fd6b-b85e-183c1a66a24b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.045025] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936264, 'name': ReconfigVM_Task, 'duration_secs': 0.595925} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.045025] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 49707589-4969-4e08-882b-2a2c94bc0d85/49707589-4969-4e08-882b-2a2c94bc0d85.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.045215] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61ca2232-bc05-4de4-b252-06511a924717 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.053625] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 996.053625] env[61985]: value = "task-936270" [ 996.053625] env[61985]: _type = "Task" [ 996.053625] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.065539] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936270, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.068377] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936267, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087946} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.068660] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.069416] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03df2e8-d15f-42ae-9579-3a86dcb6417c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.091038] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] afa46da0-3bd1-47ea-bdf0-28eed87fd627/afa46da0-3bd1-47ea-bdf0-28eed87fd627.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.091343] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f58a9f89-8844-4684-879b-cb2526009734 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.111482] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 996.111482] env[61985]: value = "task-936271" [ 996.111482] env[61985]: _type = "Task" [ 996.111482] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.119714] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936271, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.139549] env[61985]: DEBUG oslo_vmware.api [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936268, 'name': ReconfigVM_Task, 'duration_secs': 0.387032} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.141141] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Reconfigured VM instance instance-00000043 to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 996.145509] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27888f9e-7deb-4a94-988c-1cd13abd6105 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.161820] env[61985]: DEBUG oslo_vmware.api [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 996.161820] env[61985]: value = "task-936272" [ 996.161820] env[61985]: _type = "Task" [ 996.161820] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.170871] env[61985]: DEBUG oslo_vmware.api [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936272, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.353075] env[61985]: DEBUG nova.compute.utils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.357614] env[61985]: DEBUG nova.compute.manager [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 996.358339] env[61985]: DEBUG nova.network.neutron [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 996.420454] env[61985]: DEBUG nova.policy [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c25a115ea10046c9a29269373dbdabed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b5ecd127ecd40c5a980646e8cb45f6c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 996.501641] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936269, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.536995] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520e3842-55cb-fd6b-b85e-183c1a66a24b, 'name': SearchDatastore_Task, 'duration_secs': 0.01289} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.540155] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.540478] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] b6eb50ff-e685-4e8a-92f5-c2661136c361/b6eb50ff-e685-4e8a-92f5-c2661136c361.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 996.540973] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0d40bbf6-f2f7-4bcb-ba7d-8ef171ba6016 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.548105] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 996.548105] env[61985]: value = "task-936273" [ 996.548105] env[61985]: _type = "Task" [ 996.548105] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.560601] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936273, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.567016] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936270, 'name': Rename_Task, 'duration_secs': 0.352496} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.567016] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 996.568955] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b425fb44-e82d-4d92-87d2-9d4ba83f0b04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.576498] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 996.576498] env[61985]: value = "task-936274" [ 996.576498] env[61985]: _type = "Task" [ 996.576498] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.585380] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936274, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.627681] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936271, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.677938] env[61985]: DEBUG oslo_vmware.api [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936272, 'name': ReconfigVM_Task, 'duration_secs': 0.274062} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.678338] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211422', 'volume_id': '7edc010c-38ef-42c0-af8a-550bed9ea645', 'name': 'volume-7edc010c-38ef-42c0-af8a-550bed9ea645', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6dd8218b-c14d-40c9-87df-097fab06c669', 'attached_at': '', 'detached_at': '', 'volume_id': '7edc010c-38ef-42c0-af8a-550bed9ea645', 'serial': '7edc010c-38ef-42c0-af8a-550bed9ea645'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 996.712223] env[61985]: DEBUG nova.network.neutron [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Successfully created port: c38204b9-b8fd-45ce-8fb5-f1b64f8e1649 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 996.755103] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac5d506-b3df-48c5-b2ca-b33dbe38c993 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.761791] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9239e214-ded7-40b5-a86f-f2f7e46c8eca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.794347] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ed272f-3ed9-4ac8-96e7-17bc4a819211 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.803232] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9509d9c-7f27-4af5-bb15-62cf6450c1d9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.828304] env[61985]: DEBUG nova.compute.provider_tree [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.857812] env[61985]: DEBUG nova.compute.manager [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 997.005102] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936269, 'name': ReconfigVM_Task, 'duration_secs': 0.561786} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.005102] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Reconfigured VM instance instance-00000047 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 997.005788] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0644d3d2-1ae2-42db-a96a-e046f0f1515e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.042928] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] a77ca23f-b2c0-4822-8e48-3e47e0dadb27/a77ca23f-b2c0-4822-8e48-3e47e0dadb27.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.043925] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08a5ac91-689c-493b-8162-1630c67f0195 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.088021] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936273, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.088529] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 997.088529] env[61985]: value = "task-936275" [ 997.088529] env[61985]: _type = "Task" [ 997.088529] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.099480] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936274, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.104894] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936275, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.123864] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936271, 'name': ReconfigVM_Task, 'duration_secs': 0.580413} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.124251] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Reconfigured VM instance instance-00000054 to attach disk [datastore2] afa46da0-3bd1-47ea-bdf0-28eed87fd627/afa46da0-3bd1-47ea-bdf0-28eed87fd627.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.124929] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59cf94b5-a833-4996-ad7b-966c003f5892 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.134154] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 997.134154] env[61985]: value = "task-936276" [ 997.134154] env[61985]: _type = "Task" [ 997.134154] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.145031] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936276, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.231884] env[61985]: DEBUG nova.objects.instance [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lazy-loading 'flavor' on Instance uuid 6dd8218b-c14d-40c9-87df-097fab06c669 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.333159] env[61985]: DEBUG nova.scheduler.client.report [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 997.580576] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936273, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586435} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.584077] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] b6eb50ff-e685-4e8a-92f5-c2661136c361/b6eb50ff-e685-4e8a-92f5-c2661136c361.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 997.584451] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.584664] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1d58287-8dec-465d-9d02-b65ff3b12a4d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.596905] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936274, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.598815] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 997.598815] env[61985]: value = "task-936277" [ 997.598815] env[61985]: _type = "Task" [ 997.598815] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.604936] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936275, 'name': ReconfigVM_Task, 'duration_secs': 0.346163} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.605695] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Reconfigured VM instance instance-00000047 to attach disk [datastore1] a77ca23f-b2c0-4822-8e48-3e47e0dadb27/a77ca23f-b2c0-4822-8e48-3e47e0dadb27.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.606602] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69979aae-4eb3-4fe9-a97c-62db1a77ced7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.612193] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936277, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.630339] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1197a98-995f-4b97-93a8-723848b56749 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.654648] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852f02a2-6626-4e50-a644-b8d1f22fd6c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.664960] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936276, 'name': Rename_Task, 'duration_secs': 0.14849} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.665736] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 997.666081] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec4d9d84-a6a1-40b9-8251-b20a7def85d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.687764] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f279704-b6c9-4776-be36-b718445d5612 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.690908] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 997.690908] env[61985]: value = "task-936278" [ 997.690908] env[61985]: _type = "Task" [ 997.690908] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.699830] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 997.703846] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d076e17-434f-48d5-96a9-0af0c1ab3827 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.705739] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936278, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.713780] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 997.713780] env[61985]: value = "task-936279" [ 997.713780] env[61985]: _type = "Task" [ 997.713780] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.723971] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936279, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.838809] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.993s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.842117] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.936s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.843882] env[61985]: INFO nova.compute.claims [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 997.863475] env[61985]: INFO nova.scheduler.client.report [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Deleted allocations for instance 4da2d122-aa36-4cad-9521-498b53b9f2f6 [ 997.868862] env[61985]: DEBUG nova.compute.manager [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 997.896813] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 997.897102] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 997.897271] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 997.897503] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 997.897615] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 997.897791] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 997.898023] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 997.898194] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 997.898368] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 997.898655] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 997.899242] env[61985]: DEBUG nova.virt.hardware [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 997.900392] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1a8b68-c013-42f1-b4f5-92a65db23446 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.911188] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f47832d-e98d-4d2f-8f5c-0f6ff68dc563 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.094263] env[61985]: DEBUG oslo_vmware.api [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936274, 'name': PowerOnVM_Task, 'duration_secs': 1.132692} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.094708] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 998.094978] env[61985]: INFO nova.compute.manager [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Took 12.13 seconds to spawn the instance on the hypervisor. [ 998.095260] env[61985]: DEBUG nova.compute.manager [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.096425] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcf7223-6129-4793-b607-c26b080cc261 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.112574] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936277, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087543} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.114984] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.122170] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c490c296-e98f-4a75-8e60-52f654867ced {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.159368] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] b6eb50ff-e685-4e8a-92f5-c2661136c361/b6eb50ff-e685-4e8a-92f5-c2661136c361.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.161738] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4dc1c2cf-503c-4cac-9227-0f0799ea5ec9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.177185] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "625a3143-d138-4b52-aeb7-9e365c6f1128" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.177185] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.177285] env[61985]: DEBUG nova.compute.manager [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Going to confirm migration 4 {{(pid=61985) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4789}} [ 998.179751] env[61985]: DEBUG nova.compute.manager [req-e42a7e66-9d5d-4f3f-8a93-36ccfbafd940 req-5647adc6-afbe-4b9a-952b-b7085d749659 service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Received event network-vif-plugged-c38204b9-b8fd-45ce-8fb5-f1b64f8e1649 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 998.179751] env[61985]: DEBUG oslo_concurrency.lockutils [req-e42a7e66-9d5d-4f3f-8a93-36ccfbafd940 req-5647adc6-afbe-4b9a-952b-b7085d749659 service nova] Acquiring lock "bed26b5f-7ca5-405a-884e-02b2495dd977-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.180868] env[61985]: DEBUG oslo_concurrency.lockutils [req-e42a7e66-9d5d-4f3f-8a93-36ccfbafd940 req-5647adc6-afbe-4b9a-952b-b7085d749659 service nova] Lock "bed26b5f-7ca5-405a-884e-02b2495dd977-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.180868] env[61985]: DEBUG oslo_concurrency.lockutils [req-e42a7e66-9d5d-4f3f-8a93-36ccfbafd940 req-5647adc6-afbe-4b9a-952b-b7085d749659 service nova] Lock "bed26b5f-7ca5-405a-884e-02b2495dd977-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.180868] env[61985]: DEBUG nova.compute.manager [req-e42a7e66-9d5d-4f3f-8a93-36ccfbafd940 req-5647adc6-afbe-4b9a-952b-b7085d749659 service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] No waiting events found dispatching network-vif-plugged-c38204b9-b8fd-45ce-8fb5-f1b64f8e1649 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 998.180868] env[61985]: WARNING nova.compute.manager [req-e42a7e66-9d5d-4f3f-8a93-36ccfbafd940 req-5647adc6-afbe-4b9a-952b-b7085d749659 service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Received unexpected event network-vif-plugged-c38204b9-b8fd-45ce-8fb5-f1b64f8e1649 for instance with vm_state building and task_state spawning. [ 998.189069] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 998.189069] env[61985]: value = "task-936280" [ 998.189069] env[61985]: _type = "Task" [ 998.189069] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.205316] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936280, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.209659] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936278, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.223835] env[61985]: DEBUG oslo_vmware.api [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936279, 'name': PowerOnVM_Task, 'duration_secs': 0.45679} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.224179] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 998.240324] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d711fc6c-b94c-4d56-9356-fcc577e57a97 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.277s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.259613] env[61985]: DEBUG nova.network.neutron [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Successfully updated port: c38204b9-b8fd-45ce-8fb5-f1b64f8e1649 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 998.374773] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ce155643-164c-4513-88ae-fc2b952ee168 tempest-ServerMetadataTestJSON-190493831 tempest-ServerMetadataTestJSON-190493831-project-member] Lock "4da2d122-aa36-4cad-9521-498b53b9f2f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.956s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.636899] env[61985]: INFO nova.compute.manager [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Took 42.25 seconds to build instance. [ 998.703626] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936280, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.710226] env[61985]: DEBUG oslo_vmware.api [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936278, 'name': PowerOnVM_Task, 'duration_secs': 0.707794} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.710577] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 998.710796] env[61985]: INFO nova.compute.manager [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Took 10.15 seconds to spawn the instance on the hypervisor. [ 998.711027] env[61985]: DEBUG nova.compute.manager [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.711861] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8ab7e3-9d4c-4fc0-8013-50897f04d3dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.762427] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.762427] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.762427] env[61985]: DEBUG nova.network.neutron [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 998.762427] env[61985]: DEBUG nova.objects.instance [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lazy-loading 'info_cache' on Instance uuid 625a3143-d138-4b52-aeb7-9e365c6f1128 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.763668] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquiring lock "refresh_cache-bed26b5f-7ca5-405a-884e-02b2495dd977" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.763818] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquired lock "refresh_cache-bed26b5f-7ca5-405a-884e-02b2495dd977" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.763922] env[61985]: DEBUG nova.network.neutron [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.138753] env[61985]: DEBUG oslo_concurrency.lockutils [None req-52293caf-240c-4689-bf5d-3c1c7fcbee9a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "49707589-4969-4e08-882b-2a2c94bc0d85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.758s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.139616] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81cda35-5be8-45f6-9a07-66948d969d37 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.148379] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ccca06f-6be7-4ff5-8722-e6ad7259c091 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.179754] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b2b090-6195-4478-88b5-7885b5591924 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.189702] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c74cc90-64fd-436b-849e-fc9049afd009 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.205043] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936280, 'name': ReconfigVM_Task, 'duration_secs': 0.601064} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.216342] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Reconfigured VM instance instance-00000055 to attach disk [datastore1] b6eb50ff-e685-4e8a-92f5-c2661136c361/b6eb50ff-e685-4e8a-92f5-c2661136c361.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.217617] env[61985]: DEBUG nova.compute.provider_tree [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 999.222416] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ae22f16-8104-4a42-8c28-51af90c37519 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.230806] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 999.230806] env[61985]: value = "task-936281" [ 999.230806] env[61985]: _type = "Task" [ 999.230806] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.235052] env[61985]: INFO nova.compute.manager [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Took 41.11 seconds to build instance. [ 999.235767] env[61985]: INFO nova.compute.manager [None req-bb2cb2e5-b948-4fa1-8030-3af0a49e4d06 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance to original state: 'active' [ 999.251987] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936281, 'name': Rename_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.303498] env[61985]: DEBUG nova.network.neutron [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 999.446973] env[61985]: DEBUG nova.network.neutron [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Updating instance_info_cache with network_info: [{"id": "c38204b9-b8fd-45ce-8fb5-f1b64f8e1649", "address": "fa:16:3e:00:a3:72", "network": {"id": "7cacf96e-e5f0-4099-83b3-7c838a7c45f4", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-347822925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b5ecd127ecd40c5a980646e8cb45f6c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a7ba8d0-0208-4af7-af44-2a5ad382f9be", "external-id": "nsx-vlan-transportzone-742", "segmentation_id": 742, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc38204b9-b8", "ovs_interfaceid": "c38204b9-b8fd-45ce-8fb5-f1b64f8e1649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.749895] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b9a09a21-f93d-4e5b-8bc5-68a2d9cc8e0b tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.635s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.754193] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936281, 'name': Rename_Task, 'duration_secs': 0.332254} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.757374] env[61985]: ERROR nova.scheduler.client.report [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [req-631487e3-7b82-49eb-ada2-434def10517c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-631487e3-7b82-49eb-ada2-434def10517c"}]} [ 999.757808] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 999.761508] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c71f2cb3-05ec-4984-b099-672c4c25cd40 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.773522] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 999.773522] env[61985]: value = "task-936282" [ 999.773522] env[61985]: _type = "Task" [ 999.773522] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.787069] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936282, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.788173] env[61985]: DEBUG nova.scheduler.client.report [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 999.802564] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "49707589-4969-4e08-882b-2a2c94bc0d85" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.802882] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "49707589-4969-4e08-882b-2a2c94bc0d85" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.803028] env[61985]: DEBUG nova.compute.manager [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.803938] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc8873f-ac45-4d04-9e0f-e719d09c5972 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.813314] env[61985]: DEBUG nova.compute.manager [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61985) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3374}} [ 999.813314] env[61985]: DEBUG nova.objects.instance [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lazy-loading 'flavor' on Instance uuid 49707589-4969-4e08-882b-2a2c94bc0d85 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.814474] env[61985]: DEBUG nova.scheduler.client.report [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 999.814717] env[61985]: DEBUG nova.compute.provider_tree [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 999.835350] env[61985]: DEBUG nova.scheduler.client.report [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 999.847820] env[61985]: DEBUG oslo_concurrency.lockutils [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "6dd8218b-c14d-40c9-87df-097fab06c669" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.848055] env[61985]: DEBUG oslo_concurrency.lockutils [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.848523] env[61985]: DEBUG oslo_concurrency.lockutils [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "6dd8218b-c14d-40c9-87df-097fab06c669-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.848611] env[61985]: DEBUG oslo_concurrency.lockutils [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.848882] env[61985]: DEBUG oslo_concurrency.lockutils [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.851237] env[61985]: INFO nova.compute.manager [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Terminating instance [ 999.853215] env[61985]: DEBUG nova.compute.manager [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 999.853414] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 999.854304] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ee434d-f5b7-4770-a2c0-2c2d38c51b51 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.859704] env[61985]: DEBUG nova.scheduler.client.report [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 999.867488] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 999.867488] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8aef8e08-371c-46c0-b694-2e06ecd9a2b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.881976] env[61985]: DEBUG oslo_vmware.api [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 999.881976] env[61985]: value = "task-936283" [ 999.881976] env[61985]: _type = "Task" [ 999.881976] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.900754] env[61985]: DEBUG oslo_vmware.api [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936283, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.949518] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Releasing lock "refresh_cache-bed26b5f-7ca5-405a-884e-02b2495dd977" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.949956] env[61985]: DEBUG nova.compute.manager [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Instance network_info: |[{"id": "c38204b9-b8fd-45ce-8fb5-f1b64f8e1649", "address": "fa:16:3e:00:a3:72", "network": {"id": "7cacf96e-e5f0-4099-83b3-7c838a7c45f4", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-347822925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b5ecd127ecd40c5a980646e8cb45f6c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a7ba8d0-0208-4af7-af44-2a5ad382f9be", "external-id": "nsx-vlan-transportzone-742", "segmentation_id": 742, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc38204b9-b8", "ovs_interfaceid": "c38204b9-b8fd-45ce-8fb5-f1b64f8e1649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 999.951288] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:a3:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0a7ba8d0-0208-4af7-af44-2a5ad382f9be', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c38204b9-b8fd-45ce-8fb5-f1b64f8e1649', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.958763] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Creating folder: Project (3b5ecd127ecd40c5a980646e8cb45f6c). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 999.961737] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5aedef4e-1b7f-4b57-b409-bf9a3801b818 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.977965] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Created folder: Project (3b5ecd127ecd40c5a980646e8cb45f6c) in parent group-v211285. [ 999.978187] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Creating folder: Instances. Parent ref: group-v211437. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 999.978437] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aee4a7cb-7562-44e4-8a44-d5745e60f775 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.990820] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Created folder: Instances in parent group-v211437. [ 999.991092] env[61985]: DEBUG oslo.service.loopingcall [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.991300] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 999.991512] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0ba4830a-c552-483d-b9f4-f56b5fd0a957 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.023750] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1000.023750] env[61985]: value = "task-936286" [ 1000.023750] env[61985]: _type = "Task" [ 1000.023750] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.033565] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936286, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.065281] env[61985]: DEBUG nova.network.neutron [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance_info_cache with network_info: [{"id": "262813ed-9baf-4240-b57d-fc4a648b5532", "address": "fa:16:3e:d3:c1:6a", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262813ed-9b", "ovs_interfaceid": "262813ed-9baf-4240-b57d-fc4a648b5532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.278682] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c087be87-c2f4-45d2-bfde-32a841c01c94 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.300414] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b85db2-5231-4cb2-977f-872c02e8cbc0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.305262] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936282, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.339313] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0946af3-e87d-4b3f-a8af-4f0ed13b7f0f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.342903] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1000.343152] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dae4d5af-224b-4a56-aaee-11c03a597d5f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.354254] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4673ee88-659d-422f-b157-ef75fc875ae3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.359611] env[61985]: DEBUG oslo_vmware.api [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1000.359611] env[61985]: value = "task-936287" [ 1000.359611] env[61985]: _type = "Task" [ 1000.359611] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.375920] env[61985]: DEBUG nova.compute.provider_tree [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.385316] env[61985]: DEBUG oslo_vmware.api [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936287, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.396398] env[61985]: DEBUG oslo_vmware.api [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936283, 'name': PowerOffVM_Task, 'duration_secs': 0.304221} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.396685] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1000.396858] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1000.397134] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0605eaba-aaea-41bc-8ccf-6eead3d2e95e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.505778] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1000.505778] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1000.505778] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Deleting the datastore file [datastore2] 6dd8218b-c14d-40c9-87df-097fab06c669 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.507222] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d6e0207-d874-44f2-adc2-f61ade008df5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.510782] env[61985]: DEBUG nova.compute.manager [req-6feac5b7-bff5-4301-b994-cec12a29df16 req-231f88ae-225c-4ce4-84a5-c90904b7b438 service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Received event network-changed-c38204b9-b8fd-45ce-8fb5-f1b64f8e1649 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1000.511018] env[61985]: DEBUG nova.compute.manager [req-6feac5b7-bff5-4301-b994-cec12a29df16 req-231f88ae-225c-4ce4-84a5-c90904b7b438 service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Refreshing instance network info cache due to event network-changed-c38204b9-b8fd-45ce-8fb5-f1b64f8e1649. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1000.511310] env[61985]: DEBUG oslo_concurrency.lockutils [req-6feac5b7-bff5-4301-b994-cec12a29df16 req-231f88ae-225c-4ce4-84a5-c90904b7b438 service nova] Acquiring lock "refresh_cache-bed26b5f-7ca5-405a-884e-02b2495dd977" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.511403] env[61985]: DEBUG oslo_concurrency.lockutils [req-6feac5b7-bff5-4301-b994-cec12a29df16 req-231f88ae-225c-4ce4-84a5-c90904b7b438 service nova] Acquired lock "refresh_cache-bed26b5f-7ca5-405a-884e-02b2495dd977" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.511572] env[61985]: DEBUG nova.network.neutron [req-6feac5b7-bff5-4301-b994-cec12a29df16 req-231f88ae-225c-4ce4-84a5-c90904b7b438 service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Refreshing network info cache for port c38204b9-b8fd-45ce-8fb5-f1b64f8e1649 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1000.519161] env[61985]: DEBUG oslo_vmware.api [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 1000.519161] env[61985]: value = "task-936289" [ 1000.519161] env[61985]: _type = "Task" [ 1000.519161] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.531973] env[61985]: DEBUG oslo_vmware.api [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936289, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.539125] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936286, 'name': CreateVM_Task, 'duration_secs': 0.412058} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.539340] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1000.540069] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.540259] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.540577] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1000.540844] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-917c9b9e-6157-4c4f-9db7-c8f5f553eb80 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.546256] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for the task: (returnval){ [ 1000.546256] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5222987a-da16-f1f0-aff2-128648ccf9a4" [ 1000.546256] env[61985]: _type = "Task" [ 1000.546256] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.555504] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5222987a-da16-f1f0-aff2-128648ccf9a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.571211] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-625a3143-d138-4b52-aeb7-9e365c6f1128" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.571604] env[61985]: DEBUG nova.objects.instance [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lazy-loading 'migration_context' on Instance uuid 625a3143-d138-4b52-aeb7-9e365c6f1128 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.791117] env[61985]: DEBUG oslo_vmware.api [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936282, 'name': PowerOnVM_Task, 'duration_secs': 0.594026} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.791416] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1000.791627] env[61985]: INFO nova.compute.manager [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Took 9.81 seconds to spawn the instance on the hypervisor. [ 1000.791815] env[61985]: DEBUG nova.compute.manager [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.794183] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce5b75d-6c1b-4730-9a92-3fd7dd052edf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.871858] env[61985]: DEBUG oslo_vmware.api [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936287, 'name': PowerOffVM_Task, 'duration_secs': 0.201728} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.872234] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1000.872480] env[61985]: DEBUG nova.compute.manager [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.873255] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4eeb3b-d733-4df0-89c7-6c863f2655cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.920896] env[61985]: DEBUG nova.scheduler.client.report [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 115 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1000.921251] env[61985]: DEBUG nova.compute.provider_tree [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 115 to 116 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1000.921457] env[61985]: DEBUG nova.compute.provider_tree [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1001.035427] env[61985]: DEBUG oslo_vmware.api [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936289, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155914} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.035427] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.035427] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1001.035427] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1001.035427] env[61985]: INFO nova.compute.manager [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1001.035427] env[61985]: DEBUG oslo.service.loopingcall [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.035427] env[61985]: DEBUG nova.compute.manager [-] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1001.035427] env[61985]: DEBUG nova.network.neutron [-] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1001.063437] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5222987a-da16-f1f0-aff2-128648ccf9a4, 'name': SearchDatastore_Task, 'duration_secs': 0.008912} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.063931] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.064179] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1001.064431] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.064580] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.064774] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1001.065072] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-290c7e7d-6ac5-4460-8408-df90eccffc76 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.076179] env[61985]: DEBUG nova.objects.base [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Object Instance<625a3143-d138-4b52-aeb7-9e365c6f1128> lazy-loaded attributes: info_cache,migration_context {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1001.076849] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1001.076849] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1001.078031] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11edabc0-76a3-43ee-8e56-8404c8f94762 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.081773] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe70a4a0-1c1a-43f2-9843-08728ea5f66c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.089384] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for the task: (returnval){ [ 1001.089384] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f74af5-8f09-ed56-d0f4-d23452bfe2ff" [ 1001.089384] env[61985]: _type = "Task" [ 1001.089384] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.110294] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c83969a-f361-4571-b5fb-9a667c3889b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.117299] env[61985]: DEBUG oslo_vmware.api [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1001.117299] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5269f52c-b8fd-0ce2-0c65-72965dc7923e" [ 1001.117299] env[61985]: _type = "Task" [ 1001.117299] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.121235] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f74af5-8f09-ed56-d0f4-d23452bfe2ff, 'name': SearchDatastore_Task, 'duration_secs': 0.012222} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.125545] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a8c8976-fcad-4b17-9c6b-812abe1b289e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.138179] env[61985]: DEBUG oslo_vmware.api [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5269f52c-b8fd-0ce2-0c65-72965dc7923e, 'name': SearchDatastore_Task, 'duration_secs': 0.007564} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.138179] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.138179] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for the task: (returnval){ [ 1001.138179] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5244c3fc-5ef9-21a2-68dd-1b112e799f37" [ 1001.138179] env[61985]: _type = "Task" [ 1001.138179] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.147365] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5244c3fc-5ef9-21a2-68dd-1b112e799f37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.214917] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.215234] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.215419] env[61985]: INFO nova.compute.manager [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Shelving [ 1001.313135] env[61985]: INFO nova.compute.manager [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Took 37.48 seconds to build instance. [ 1001.392332] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb265b1f-fce6-405a-adbc-5602c18f1528 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "49707589-4969-4e08-882b-2a2c94bc0d85" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.589s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.427699] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.586s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.428591] env[61985]: DEBUG nova.compute.manager [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1001.433070] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.767s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.435176] env[61985]: INFO nova.compute.claims [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.525805] env[61985]: DEBUG nova.network.neutron [req-6feac5b7-bff5-4301-b994-cec12a29df16 req-231f88ae-225c-4ce4-84a5-c90904b7b438 service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Updated VIF entry in instance network info cache for port c38204b9-b8fd-45ce-8fb5-f1b64f8e1649. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1001.526637] env[61985]: DEBUG nova.network.neutron [req-6feac5b7-bff5-4301-b994-cec12a29df16 req-231f88ae-225c-4ce4-84a5-c90904b7b438 service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Updating instance_info_cache with network_info: [{"id": "c38204b9-b8fd-45ce-8fb5-f1b64f8e1649", "address": "fa:16:3e:00:a3:72", "network": {"id": "7cacf96e-e5f0-4099-83b3-7c838a7c45f4", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-347822925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b5ecd127ecd40c5a980646e8cb45f6c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a7ba8d0-0208-4af7-af44-2a5ad382f9be", "external-id": "nsx-vlan-transportzone-742", "segmentation_id": 742, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc38204b9-b8", "ovs_interfaceid": "c38204b9-b8fd-45ce-8fb5-f1b64f8e1649", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.651589] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5244c3fc-5ef9-21a2-68dd-1b112e799f37, 'name': SearchDatastore_Task, 'duration_secs': 0.010439} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.651745] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.652057] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] bed26b5f-7ca5-405a-884e-02b2495dd977/bed26b5f-7ca5-405a-884e-02b2495dd977.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1001.652333] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88af5c21-ac9e-4305-9f6b-730dbb8d3205 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.661571] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for the task: (returnval){ [ 1001.661571] env[61985]: value = "task-936290" [ 1001.661571] env[61985]: _type = "Task" [ 1001.661571] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.680850] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936290, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.724039] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1001.724323] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9591d06e-86ce-4cc9-8e82-1d843e43b53c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.733922] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1001.733922] env[61985]: value = "task-936291" [ 1001.733922] env[61985]: _type = "Task" [ 1001.733922] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.744032] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936291, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.815918] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6f228451-4e1b-4331-9581-4dc1cb17850f tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "b6eb50ff-e685-4e8a-92f5-c2661136c361" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.991s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.945268] env[61985]: DEBUG nova.compute.utils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1001.954639] env[61985]: DEBUG nova.compute.manager [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1001.955024] env[61985]: DEBUG nova.network.neutron [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1002.028842] env[61985]: DEBUG nova.policy [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '905fa37e1f784a56b3edb9818b03d685', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23a4809e353a4ecaa1bd53612fb4afc5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1002.033409] env[61985]: DEBUG oslo_concurrency.lockutils [req-6feac5b7-bff5-4301-b994-cec12a29df16 req-231f88ae-225c-4ce4-84a5-c90904b7b438 service nova] Releasing lock "refresh_cache-bed26b5f-7ca5-405a-884e-02b2495dd977" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.174290] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.174290] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.174524] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.175242] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.175426] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.180607] env[61985]: INFO nova.compute.manager [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Terminating instance [ 1002.185999] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936290, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.186774] env[61985]: DEBUG nova.compute.manager [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1002.187072] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1002.190698] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e93782-358f-4d21-b720-7e728c365538 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.203322] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1002.204144] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-599146a8-fbdc-4663-b365-9483e7091e01 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.220494] env[61985]: DEBUG oslo_vmware.api [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 1002.220494] env[61985]: value = "task-936292" [ 1002.220494] env[61985]: _type = "Task" [ 1002.220494] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.238955] env[61985]: DEBUG oslo_vmware.api [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936292, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.256933] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936291, 'name': PowerOffVM_Task, 'duration_secs': 0.303484} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.257208] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1002.258146] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6384492-251d-480a-aab4-8cdd4a8d5f1a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.289166] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac3eebb-ff5e-409e-8a1a-e158d810fe8b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.400642] env[61985]: DEBUG nova.network.neutron [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Successfully created port: e35f2970-0b73-4b31-925f-56ae8e0e7f39 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1002.419577] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9791f908-d022-486b-82b3-8a9521e0a8b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.430207] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1972d83-17e1-4414-af5c-3eccceb606e4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.464651] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd8278c-9157-4ca9-ac1e-0e0378a015e3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.472402] env[61985]: DEBUG nova.compute.manager [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1002.479152] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4bf12ef-2427-43f1-8c38-86d411e9930b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.494424] env[61985]: DEBUG nova.compute.provider_tree [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.673846] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936290, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612157} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.674215] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] bed26b5f-7ca5-405a-884e-02b2495dd977/bed26b5f-7ca5-405a-884e-02b2495dd977.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1002.674465] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1002.674845] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88e8ba98-4c3d-4fbd-865f-77a5d74a5388 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.683077] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for the task: (returnval){ [ 1002.683077] env[61985]: value = "task-936293" [ 1002.683077] env[61985]: _type = "Task" [ 1002.683077] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.693978] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.720296] env[61985]: DEBUG nova.compute.manager [req-ca857580-e619-4a6c-bdce-aaadf16b43ab req-94eb3e2b-dea4-45c9-a2a2-fdada2e8c650 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Received event network-vif-deleted-97f641fb-39ac-441c-bacc-6705d1ea6e98 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1002.720430] env[61985]: INFO nova.compute.manager [req-ca857580-e619-4a6c-bdce-aaadf16b43ab req-94eb3e2b-dea4-45c9-a2a2-fdada2e8c650 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Neutron deleted interface 97f641fb-39ac-441c-bacc-6705d1ea6e98; detaching it from the instance and deleting it from the info cache [ 1002.720618] env[61985]: DEBUG nova.network.neutron [req-ca857580-e619-4a6c-bdce-aaadf16b43ab req-94eb3e2b-dea4-45c9-a2a2-fdada2e8c650 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.735065] env[61985]: DEBUG oslo_vmware.api [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936292, 'name': PowerOffVM_Task, 'duration_secs': 0.283352} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.735065] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1002.735065] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1002.735434] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1450f8f-20b0-4c56-8120-44ec2c5f692d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.748892] env[61985]: DEBUG nova.network.neutron [-] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.802123] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1002.802475] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6ff1291d-e9c3-45c0-97bc-260d8c7aac97 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.811444] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1002.811444] env[61985]: value = "task-936295" [ 1002.811444] env[61985]: _type = "Task" [ 1002.811444] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.824087] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936295, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.828646] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1002.831222] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1002.831222] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Deleting the datastore file [datastore1] a77ca23f-b2c0-4822-8e48-3e47e0dadb27 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.831222] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e2a4f82-3928-4584-b0e6-feb574bb9be6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.837621] env[61985]: DEBUG oslo_vmware.api [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 1002.837621] env[61985]: value = "task-936296" [ 1002.837621] env[61985]: _type = "Task" [ 1002.837621] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.850151] env[61985]: DEBUG oslo_vmware.api [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936296, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.998670] env[61985]: DEBUG nova.scheduler.client.report [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.197503] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092312} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.198135] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.198990] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1f466d-e103-47ff-9c93-e5b79b3dc07f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.226563] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] bed26b5f-7ca5-405a-884e-02b2495dd977/bed26b5f-7ca5-405a-884e-02b2495dd977.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.228286] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-639af777-5438-4295-a2e7-d2633ad15ee4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.246550] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7652f7d0-8772-48ec-a899-94f7ce8b914e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.252220] env[61985]: INFO nova.compute.manager [-] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Took 2.22 seconds to deallocate network for instance. [ 1003.263833] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for the task: (returnval){ [ 1003.263833] env[61985]: value = "task-936297" [ 1003.263833] env[61985]: _type = "Task" [ 1003.263833] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.267291] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8252423e-a5c7-44dd-9502-f41ccb7adb5f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.288248] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936297, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.325271] env[61985]: DEBUG nova.compute.manager [req-ca857580-e619-4a6c-bdce-aaadf16b43ab req-94eb3e2b-dea4-45c9-a2a2-fdada2e8c650 service nova] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Detach interface failed, port_id=97f641fb-39ac-441c-bacc-6705d1ea6e98, reason: Instance 6dd8218b-c14d-40c9-87df-097fab06c669 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1003.339845] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936295, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.350967] env[61985]: DEBUG oslo_vmware.api [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936296, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.288804} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.351257] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.351450] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1003.351639] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1003.351820] env[61985]: INFO nova.compute.manager [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1003.352104] env[61985]: DEBUG oslo.service.loopingcall [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.353116] env[61985]: DEBUG nova.compute.manager [-] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1003.353116] env[61985]: DEBUG nova.network.neutron [-] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1003.488327] env[61985]: DEBUG nova.compute.manager [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1003.512355] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.079s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.512880] env[61985]: DEBUG nova.compute.manager [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1003.522756] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1003.522998] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1003.527375] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.527699] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1003.527835] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.527995] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1003.528326] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1003.528422] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1003.528630] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1003.528861] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1003.529073] env[61985]: DEBUG nova.virt.hardware [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1003.529445] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.758s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.529642] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.534671] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.512s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.538761] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0bdfda-5da1-464f-b00c-f96d9e30a2f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.556028] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04741384-8ca1-486f-8ca4-239d09a566ad {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.583740] env[61985]: INFO nova.scheduler.client.report [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleted allocations for instance 2322ab72-9841-41fb-9d60-2812baabe108 [ 1003.688441] env[61985]: INFO nova.compute.manager [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Rebuilding instance [ 1003.746783] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "b6eb50ff-e685-4e8a-92f5-c2661136c361" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.747755] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "b6eb50ff-e685-4e8a-92f5-c2661136c361" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.748271] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "b6eb50ff-e685-4e8a-92f5-c2661136c361-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.748661] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "b6eb50ff-e685-4e8a-92f5-c2661136c361-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.749055] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "b6eb50ff-e685-4e8a-92f5-c2661136c361-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.752043] env[61985]: DEBUG nova.compute.manager [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.753155] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2eac685-b024-4a08-9860-cd0d9da2a97a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.757813] env[61985]: INFO nova.compute.manager [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Terminating instance [ 1003.760832] env[61985]: DEBUG nova.compute.manager [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1003.761014] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1003.762303] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4430d9-3c4d-4ffc-a814-0abc6fb97dc0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.777216] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1003.779037] env[61985]: DEBUG oslo_concurrency.lockutils [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.784587] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8831e8cc-68f5-4423-8e93-240eaf9db9e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.795672] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936297, 'name': ReconfigVM_Task, 'duration_secs': 0.315596} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.796291] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Reconfigured VM instance instance-00000056 to attach disk [datastore1] bed26b5f-7ca5-405a-884e-02b2495dd977/bed26b5f-7ca5-405a-884e-02b2495dd977.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.797394] env[61985]: DEBUG oslo_vmware.api [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1003.797394] env[61985]: value = "task-936298" [ 1003.797394] env[61985]: _type = "Task" [ 1003.797394] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.797635] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c110f25-895a-4465-829c-116086c0c20f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.804926] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.804926] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.804926] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.805222] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.805722] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.810749] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for the task: (returnval){ [ 1003.810749] env[61985]: value = "task-936299" [ 1003.810749] env[61985]: _type = "Task" [ 1003.810749] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.815746] env[61985]: INFO nova.compute.manager [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Terminating instance [ 1003.817860] env[61985]: DEBUG oslo_vmware.api [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936298, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.822046] env[61985]: DEBUG nova.compute.manager [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1003.822313] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1003.823313] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5ac601-9e50-41ef-9588-08797adb63bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.839469] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936299, 'name': Rename_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.841939] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1003.843294] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-565cc00a-0031-4e5e-b79b-d8688ef7fe07 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.852250] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936295, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.854215] env[61985]: DEBUG oslo_vmware.api [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1003.854215] env[61985]: value = "task-936300" [ 1003.854215] env[61985]: _type = "Task" [ 1003.854215] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.864652] env[61985]: DEBUG oslo_vmware.api [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936300, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.047380] env[61985]: DEBUG nova.compute.utils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1004.063044] env[61985]: DEBUG nova.compute.manager [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1004.063044] env[61985]: DEBUG nova.network.neutron [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1004.085730] env[61985]: DEBUG nova.network.neutron [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Successfully updated port: e35f2970-0b73-4b31-925f-56ae8e0e7f39 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1004.092190] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6639aff3-31dd-4550-bdf1-ea2122c990ec tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "2322ab72-9841-41fb-9d60-2812baabe108" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.533s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.167804] env[61985]: DEBUG nova.policy [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '149807bc94cc4cddb59e4dad0bf1099c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8015ef21eaf4a0bb19063e261064cdc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.249509] env[61985]: DEBUG nova.compute.manager [req-3caf2c4f-7f17-4b6c-b3a3-53d3cfa20fa7 req-a4bf7923-6625-4c52-971d-f5a8b04e3439 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Received event network-vif-plugged-e35f2970-0b73-4b31-925f-56ae8e0e7f39 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1004.249747] env[61985]: DEBUG oslo_concurrency.lockutils [req-3caf2c4f-7f17-4b6c-b3a3-53d3cfa20fa7 req-a4bf7923-6625-4c52-971d-f5a8b04e3439 service nova] Acquiring lock "44361f7b-4609-476f-b4a9-58a7851e6e92-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.249988] env[61985]: DEBUG oslo_concurrency.lockutils [req-3caf2c4f-7f17-4b6c-b3a3-53d3cfa20fa7 req-a4bf7923-6625-4c52-971d-f5a8b04e3439 service nova] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.250198] env[61985]: DEBUG oslo_concurrency.lockutils [req-3caf2c4f-7f17-4b6c-b3a3-53d3cfa20fa7 req-a4bf7923-6625-4c52-971d-f5a8b04e3439 service nova] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.250379] env[61985]: DEBUG nova.compute.manager [req-3caf2c4f-7f17-4b6c-b3a3-53d3cfa20fa7 req-a4bf7923-6625-4c52-971d-f5a8b04e3439 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] No waiting events found dispatching network-vif-plugged-e35f2970-0b73-4b31-925f-56ae8e0e7f39 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.250549] env[61985]: WARNING nova.compute.manager [req-3caf2c4f-7f17-4b6c-b3a3-53d3cfa20fa7 req-a4bf7923-6625-4c52-971d-f5a8b04e3439 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Received unexpected event network-vif-plugged-e35f2970-0b73-4b31-925f-56ae8e0e7f39 for instance with vm_state building and task_state spawning. [ 1004.272818] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.273142] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02fcedb2-7d5d-4048-a6f2-24ae3983a5e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.282185] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1004.282185] env[61985]: value = "task-936301" [ 1004.282185] env[61985]: _type = "Task" [ 1004.282185] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.293568] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1004.293821] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1004.294827] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d17dfd7-4ed7-45fd-9ad4-8f290ad0016b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.305390] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1004.306169] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a8da147-cc92-499a-ae34-7d7d461f34e3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.311704] env[61985]: DEBUG oslo_vmware.api [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936298, 'name': PowerOffVM_Task, 'duration_secs': 0.211387} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.311975] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1004.312194] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1004.312454] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-844aa2e1-61e4-4684-a5bd-d1102f154ad4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.327826] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936299, 'name': Rename_Task, 'duration_secs': 0.20543} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.328573] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1004.328893] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2afa8f43-6c15-417b-bee4-9db555c750ce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.340253] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936295, 'name': CreateSnapshot_Task, 'duration_secs': 1.055294} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.341673] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1004.342248] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for the task: (returnval){ [ 1004.342248] env[61985]: value = "task-936304" [ 1004.342248] env[61985]: _type = "Task" [ 1004.342248] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.342711] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf7ae01-f951-4517-9ac8-1cb1f69f68ec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.353390] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936304, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.369781] env[61985]: DEBUG oslo_vmware.api [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936300, 'name': PowerOffVM_Task, 'duration_secs': 0.19349} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.370083] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1004.370263] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1004.370554] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a14eced8-4c9a-4010-a1ec-425aaf3f5511 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.421237] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1004.421787] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1004.421787] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleting the datastore file [datastore1] b6eb50ff-e685-4e8a-92f5-c2661136c361 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.421942] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a442812-3a12-4849-9a0c-045124806929 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.430268] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1004.430484] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1004.430667] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleting the datastore file [datastore2] 49707589-4969-4e08-882b-2a2c94bc0d85 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.430927] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-659ebda4-7a68-485b-ab37-6baac1932c8d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.438580] env[61985]: DEBUG oslo_vmware.api [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1004.438580] env[61985]: value = "task-936306" [ 1004.438580] env[61985]: _type = "Task" [ 1004.438580] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.446847] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1004.446847] env[61985]: value = "task-936307" [ 1004.446847] env[61985]: _type = "Task" [ 1004.446847] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.453947] env[61985]: DEBUG oslo_vmware.api [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936306, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.457141] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1004.457398] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1004.457661] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleting the datastore file [datastore2] 172647b8-1f1d-49cb-a0fd-63078bc1ae85 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.458069] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e7f6f31b-a207-44ba-8d5c-bc8ac8fdc242 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.463213] env[61985]: DEBUG nova.network.neutron [-] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.464463] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.471684] env[61985]: DEBUG oslo_vmware.api [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1004.471684] env[61985]: value = "task-936308" [ 1004.471684] env[61985]: _type = "Task" [ 1004.471684] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.481744] env[61985]: DEBUG oslo_vmware.api [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936308, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.561232] env[61985]: DEBUG nova.compute.manager [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1004.567684] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Applying migration context for instance 625a3143-d138-4b52-aeb7-9e365c6f1128 as it has an incoming, in-progress migration 1bf48126-5cd8-40dd-bbe0-d7c7e1d54f35. Migration status is confirming {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1004.570411] env[61985]: INFO nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating resource usage from migration 1bf48126-5cd8-40dd-bbe0-d7c7e1d54f35 [ 1004.587449] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.587584] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.588302] env[61985]: DEBUG nova.network.neutron [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1004.600514] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 761d5954-a5ca-4459-a1d6-bfc59b284bf4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.600514] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance cfd59b61-cca9-48d5-85e1-1f45d13f1e88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.600514] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 59e742d9-c51a-47dd-9ef2-2f070eae94c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.600514] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance ce757dda-f58a-47b3-b319-e6b01f05b20b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.603037] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance fea0cf39-e851-409f-86f5-31cc128a44dc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.603370] env[61985]: WARNING nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 6dd8218b-c14d-40c9-87df-097fab06c669 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1004.603517] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 073a7668-39e6-480d-9350-835a0282b456 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.603641] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 986c576e-be02-48ac-b24c-72edccab25be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.604638] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 87006fbf-ea90-4d9a-88af-001de424ac14 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.604638] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance a77ca23f-b2c0-4822-8e48-3e47e0dadb27 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.604638] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 172647b8-1f1d-49cb-a0fd-63078bc1ae85 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.604638] env[61985]: WARNING nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 942e0f09-4d34-4aa2-8a60-4c12f899712c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1004.604638] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 5f2aa808-8d1e-471e-9ef7-0f91590ec546 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.604638] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Migration 1bf48126-5cd8-40dd-bbe0-d7c7e1d54f35 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1004.604638] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 625a3143-d138-4b52-aeb7-9e365c6f1128 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.604638] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 49707589-4969-4e08-882b-2a2c94bc0d85 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.604638] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance afa46da0-3bd1-47ea-bdf0-28eed87fd627 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.605870] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance b6eb50ff-e685-4e8a-92f5-c2661136c361 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.605870] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance bed26b5f-7ca5-405a-884e-02b2495dd977 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.605870] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 44361f7b-4609-476f-b4a9-58a7851e6e92 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.605870] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 90ac7d3e-aef3-4a50-963a-606e7c58f446 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1004.743065] env[61985]: DEBUG nova.network.neutron [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Successfully created port: 63023688-9d9e-4fab-91af-9e9171bccc16 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1004.858258] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936304, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.870925] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1004.871293] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5f798fd6-19d5-48e6-8d61-5e4c1bcb6e3c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.880131] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1004.880131] env[61985]: value = "task-936309" [ 1004.880131] env[61985]: _type = "Task" [ 1004.880131] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.889223] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936309, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.950640] env[61985]: DEBUG oslo_vmware.api [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936306, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227287} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.954335] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.954470] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1004.954649] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1004.954959] env[61985]: INFO nova.compute.manager [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1004.955231] env[61985]: DEBUG oslo.service.loopingcall [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.955575] env[61985]: DEBUG nova.compute.manager [-] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.955575] env[61985]: DEBUG nova.network.neutron [-] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1004.962830] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.966400] env[61985]: INFO nova.compute.manager [-] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Took 1.61 seconds to deallocate network for instance. [ 1004.983642] env[61985]: DEBUG oslo_vmware.api [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936308, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.392387} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.984025] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.984294] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1004.984540] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1004.984820] env[61985]: INFO nova.compute.manager [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1004.985397] env[61985]: DEBUG oslo.service.loopingcall [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.985715] env[61985]: DEBUG nova.compute.manager [-] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.985924] env[61985]: DEBUG nova.network.neutron [-] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1005.110868] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8db82979-e527-4c63-8ce8-fe1456f7708c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1005.111215] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 19 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1005.111398] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4288MB phys_disk=200GB used_disk=19GB total_vcpus=48 used_vcpus=19 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1005.141113] env[61985]: DEBUG nova.network.neutron [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1005.248157] env[61985]: DEBUG nova.compute.manager [req-1a92012a-6082-4b16-a213-a0e2e8279488 req-ab3bfba9-b8a4-44ab-a998-d888f28b9385 service nova] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Received event network-vif-deleted-1d016adf-58f0-4fc2-a9d5-4e7232356db4 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1005.346388] env[61985]: DEBUG nova.network.neutron [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance_info_cache with network_info: [{"id": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "address": "fa:16:3e:13:20:58", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35f2970-0b", "ovs_interfaceid": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.363027] env[61985]: DEBUG oslo_vmware.api [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936304, 'name': PowerOnVM_Task, 'duration_secs': 0.556423} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.364263] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1005.364263] env[61985]: INFO nova.compute.manager [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Took 7.49 seconds to spawn the instance on the hypervisor. [ 1005.364263] env[61985]: DEBUG nova.compute.manager [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.364670] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05bec0a7-45a6-4231-81e5-d395b430e0ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.391661] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936309, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.457288] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d091344-d0ea-463a-bc76-1bd2e8f32811 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.469221] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd282a79-01e7-4fa3-a434-a94294a78afb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.474082] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.506271] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.507625] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a24ff8f-3271-4d4d-81e9-0cd4df240bfd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.515673] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef948386-dfab-4504-8873-876a4db02879 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.530264] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.574230] env[61985]: DEBUG nova.compute.manager [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1005.607055] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.607379] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.607564] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.607850] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.608026] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.608177] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.608391] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.608560] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.608829] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.609028] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.609208] env[61985]: DEBUG nova.virt.hardware [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.611556] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e892ed8-d17e-4432-834f-a9654a455ab2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.619023] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2e0f5e-f971-4f4d-a60b-e8b0c4f85dac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.808887] env[61985]: DEBUG nova.network.neutron [-] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.854754] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.855094] env[61985]: DEBUG nova.compute.manager [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Instance network_info: |[{"id": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "address": "fa:16:3e:13:20:58", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35f2970-0b", "ovs_interfaceid": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1005.855552] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:20:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e35f2970-0b73-4b31-925f-56ae8e0e7f39', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1005.863847] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Creating folder: Project (23a4809e353a4ecaa1bd53612fb4afc5). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1005.864314] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74e99582-075f-4509-8a36-b1558ab455ed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.876722] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Created folder: Project (23a4809e353a4ecaa1bd53612fb4afc5) in parent group-v211285. [ 1005.876921] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Creating folder: Instances. Parent ref: group-v211442. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1005.881204] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9efba42c-1c86-4add-813f-d10c20a005c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.889823] env[61985]: INFO nova.compute.manager [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Took 34.28 seconds to build instance. [ 1005.894717] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936309, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.896371] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Created folder: Instances in parent group-v211442. [ 1005.896714] env[61985]: DEBUG oslo.service.loopingcall [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.896961] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1005.897240] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-87e2c3dd-ade9-485f-b1c7-8a9acac83e20 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.917850] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1005.917850] env[61985]: value = "task-936312" [ 1005.917850] env[61985]: _type = "Task" [ 1005.917850] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.930011] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936312, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.964309] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936307, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.213917} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.964602] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1005.964813] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1005.964995] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1006.051478] env[61985]: ERROR nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [req-7ddd243a-eb52-4e59-90e0-d195fa28697c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7ddd243a-eb52-4e59-90e0-d195fa28697c"}]} [ 1006.076760] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1006.100031] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1006.100031] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.111532] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1006.131490] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1006.200312] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "885a36d9-24c6-407a-9d6b-01cd97e90674" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.201247] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.311611] env[61985]: INFO nova.compute.manager [-] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Took 1.36 seconds to deallocate network for instance. [ 1006.392936] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936309, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.395820] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ba550161-fdec-40a8-9c6d-753953f5673e tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lock "bed26b5f-7ca5-405a-884e-02b2495dd977" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.794s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.427779] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936312, 'name': CreateVM_Task, 'duration_secs': 0.340731} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.427779] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1006.428429] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.428604] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.428986] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1006.429272] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d1caa60-60d9-4ba8-85c6-7796a7ec80c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.437230] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1006.437230] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5253e5aa-005f-0ea0-bd8e-fb8f60414a23" [ 1006.437230] env[61985]: _type = "Task" [ 1006.437230] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.447444] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5253e5aa-005f-0ea0-bd8e-fb8f60414a23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.484902] env[61985]: DEBUG nova.compute.manager [req-cfb6b9fd-eaaa-493c-af53-1c91b91dc8ac req-7742fdf5-09b5-426e-afef-7860b68d2ef7 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Received event network-changed-e35f2970-0b73-4b31-925f-56ae8e0e7f39 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1006.484902] env[61985]: DEBUG nova.compute.manager [req-cfb6b9fd-eaaa-493c-af53-1c91b91dc8ac req-7742fdf5-09b5-426e-afef-7860b68d2ef7 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Refreshing instance network info cache due to event network-changed-e35f2970-0b73-4b31-925f-56ae8e0e7f39. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1006.485253] env[61985]: DEBUG oslo_concurrency.lockutils [req-cfb6b9fd-eaaa-493c-af53-1c91b91dc8ac req-7742fdf5-09b5-426e-afef-7860b68d2ef7 service nova] Acquiring lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.485253] env[61985]: DEBUG oslo_concurrency.lockutils [req-cfb6b9fd-eaaa-493c-af53-1c91b91dc8ac req-7742fdf5-09b5-426e-afef-7860b68d2ef7 service nova] Acquired lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.485334] env[61985]: DEBUG nova.network.neutron [req-cfb6b9fd-eaaa-493c-af53-1c91b91dc8ac req-7742fdf5-09b5-426e-afef-7860b68d2ef7 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Refreshing network info cache for port e35f2970-0b73-4b31-925f-56ae8e0e7f39 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1006.497772] env[61985]: DEBUG nova.network.neutron [-] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.509651] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e15fdf3-8ba4-4740-b3df-4757d81d536f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.518056] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b21152b-2a5a-455b-9b87-6f860316d97a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.550163] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0756d65d-4abd-432d-b208-cab7e7e97b70 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.558438] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67dec05c-2e15-4859-93d7-5434dd41a629 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.571674] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.703877] env[61985]: DEBUG nova.compute.manager [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1006.820582] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.893356] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936309, 'name': CloneVM_Task, 'duration_secs': 1.580176} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.893695] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Created linked-clone VM from snapshot [ 1006.894694] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2470c18-a416-4d31-9082-de33015fcb02 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.903921] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Uploading image 10a07ac1-b198-4338-8b14-ed1bbddb0250 {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1006.928774] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1006.928774] env[61985]: value = "vm-211441" [ 1006.928774] env[61985]: _type = "VirtualMachine" [ 1006.928774] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1006.930036] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d73c6e3c-843e-4771-92be-4490831da4f4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.937043] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lease: (returnval){ [ 1006.937043] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c29a5c-5ab4-7ac2-8aea-b6b3d66fbb83" [ 1006.937043] env[61985]: _type = "HttpNfcLease" [ 1006.937043] env[61985]: } obtained for exporting VM: (result){ [ 1006.937043] env[61985]: value = "vm-211441" [ 1006.937043] env[61985]: _type = "VirtualMachine" [ 1006.937043] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1006.937281] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the lease: (returnval){ [ 1006.937281] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c29a5c-5ab4-7ac2-8aea-b6b3d66fbb83" [ 1006.937281] env[61985]: _type = "HttpNfcLease" [ 1006.937281] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1006.953025] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5253e5aa-005f-0ea0-bd8e-fb8f60414a23, 'name': SearchDatastore_Task, 'duration_secs': 0.012233} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.953025] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.953025] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1006.953025] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.953300] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.953300] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1006.953559] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1006.953559] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c29a5c-5ab4-7ac2-8aea-b6b3d66fbb83" [ 1006.953559] env[61985]: _type = "HttpNfcLease" [ 1006.953559] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1006.953677] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1454340-c0c8-4dae-ad7c-f926ecadb761 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.955543] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1006.955543] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c29a5c-5ab4-7ac2-8aea-b6b3d66fbb83" [ 1006.955543] env[61985]: _type = "HttpNfcLease" [ 1006.955543] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1006.956553] env[61985]: DEBUG nova.network.neutron [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Successfully updated port: 63023688-9d9e-4fab-91af-9e9171bccc16 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.957742] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d545a17-403f-416b-b322-8d146d5609c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.966909] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525fe5ee-a9ce-6e08-8184-5b3421e34878/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1006.966909] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525fe5ee-a9ce-6e08-8184-5b3421e34878/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1006.969232] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1006.969232] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1006.970679] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52a26a5c-7e00-4494-ad00-61b56234afa4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.033203] env[61985]: INFO nova.compute.manager [-] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Took 2.05 seconds to deallocate network for instance. [ 1007.038578] env[61985]: DEBUG nova.compute.manager [req-463a3411-3ae9-4a7e-82b3-8d76fc7cf91a req-73904465-362a-4ae6-82b6-1e6ed89829dd service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Received event network-vif-plugged-63023688-9d9e-4fab-91af-9e9171bccc16 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1007.039829] env[61985]: DEBUG oslo_concurrency.lockutils [req-463a3411-3ae9-4a7e-82b3-8d76fc7cf91a req-73904465-362a-4ae6-82b6-1e6ed89829dd service nova] Acquiring lock "90ac7d3e-aef3-4a50-963a-606e7c58f446-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.039829] env[61985]: DEBUG oslo_concurrency.lockutils [req-463a3411-3ae9-4a7e-82b3-8d76fc7cf91a req-73904465-362a-4ae6-82b6-1e6ed89829dd service nova] Lock "90ac7d3e-aef3-4a50-963a-606e7c58f446-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.039829] env[61985]: DEBUG oslo_concurrency.lockutils [req-463a3411-3ae9-4a7e-82b3-8d76fc7cf91a req-73904465-362a-4ae6-82b6-1e6ed89829dd service nova] Lock "90ac7d3e-aef3-4a50-963a-606e7c58f446-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.039829] env[61985]: DEBUG nova.compute.manager [req-463a3411-3ae9-4a7e-82b3-8d76fc7cf91a req-73904465-362a-4ae6-82b6-1e6ed89829dd service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] No waiting events found dispatching network-vif-plugged-63023688-9d9e-4fab-91af-9e9171bccc16 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1007.039829] env[61985]: WARNING nova.compute.manager [req-463a3411-3ae9-4a7e-82b3-8d76fc7cf91a req-73904465-362a-4ae6-82b6-1e6ed89829dd service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Received unexpected event network-vif-plugged-63023688-9d9e-4fab-91af-9e9171bccc16 for instance with vm_state building and task_state spawning. [ 1007.048020] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1007.048020] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524371f5-f38b-3ac0-34dc-0062d0d5d421" [ 1007.048020] env[61985]: _type = "Task" [ 1007.048020] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.055531] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524371f5-f38b-3ac0-34dc-0062d0d5d421, 'name': SearchDatastore_Task, 'duration_secs': 0.012673} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.055531] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74269ec0-dee5-495e-9a17-ca8fd3a6d1e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.062941] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1007.062941] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5291bc7a-2697-34be-47d2-9a7fbc467f4b" [ 1007.062941] env[61985]: _type = "Task" [ 1007.062941] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.070995] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.071260] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.071459] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.071606] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.071796] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.071954] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.072188] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.072352] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.072524] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.072693] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.072942] env[61985]: DEBUG nova.virt.hardware [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.076522] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8d7137af-21f0-4601-ae32-42dd11011cb6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.078954] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-112698a2-96aa-4cbb-86d0-a870975f4da3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.084335] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5291bc7a-2697-34be-47d2-9a7fbc467f4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.090885] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b30a945-b076-4cb8-b229-2b241c10bbd1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.108738] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e5:10:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b48329e-7fd7-4641-845a-68423e22587b', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1007.116082] env[61985]: DEBUG oslo.service.loopingcall [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.119342] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 117 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1007.119570] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 117 to 118 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1007.119692] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1007.124542] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1007.125609] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc3d5b75-ab21-421f-92f6-a60e050158c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.151529] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1007.151529] env[61985]: value = "task-936314" [ 1007.151529] env[61985]: _type = "Task" [ 1007.151529] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.159203] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936314, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.224512] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.462013] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "refresh_cache-90ac7d3e-aef3-4a50-963a-606e7c58f446" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.462195] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired lock "refresh_cache-90ac7d3e-aef3-4a50-963a-606e7c58f446" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.462416] env[61985]: DEBUG nova.network.neutron [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1007.478129] env[61985]: DEBUG nova.compute.manager [req-2886dc82-f7ce-45ae-bdee-59a22fb75b5d req-5e1c2b53-37a8-4ab1-9c64-8c33f9b85ec1 service nova] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Received event network-vif-deleted-f33869c6-9fa7-44ab-8de2-ae92191b1096 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1007.479040] env[61985]: DEBUG nova.compute.manager [req-2886dc82-f7ce-45ae-bdee-59a22fb75b5d req-5e1c2b53-37a8-4ab1-9c64-8c33f9b85ec1 service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Received event network-changed-63023688-9d9e-4fab-91af-9e9171bccc16 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1007.479215] env[61985]: DEBUG nova.compute.manager [req-2886dc82-f7ce-45ae-bdee-59a22fb75b5d req-5e1c2b53-37a8-4ab1-9c64-8c33f9b85ec1 service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Refreshing instance network info cache due to event network-changed-63023688-9d9e-4fab-91af-9e9171bccc16. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1007.479852] env[61985]: DEBUG oslo_concurrency.lockutils [req-2886dc82-f7ce-45ae-bdee-59a22fb75b5d req-5e1c2b53-37a8-4ab1-9c64-8c33f9b85ec1 service nova] Acquiring lock "refresh_cache-90ac7d3e-aef3-4a50-963a-606e7c58f446" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.480352] env[61985]: DEBUG nova.network.neutron [req-cfb6b9fd-eaaa-493c-af53-1c91b91dc8ac req-7742fdf5-09b5-426e-afef-7860b68d2ef7 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updated VIF entry in instance network info cache for port e35f2970-0b73-4b31-925f-56ae8e0e7f39. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1007.480863] env[61985]: DEBUG nova.network.neutron [req-cfb6b9fd-eaaa-493c-af53-1c91b91dc8ac req-7742fdf5-09b5-426e-afef-7860b68d2ef7 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance_info_cache with network_info: [{"id": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "address": "fa:16:3e:13:20:58", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35f2970-0b", "ovs_interfaceid": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.547336] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.577712] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5291bc7a-2697-34be-47d2-9a7fbc467f4b, 'name': SearchDatastore_Task, 'duration_secs': 0.009961} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.578372] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.578808] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 44361f7b-4609-476f-b4a9-58a7851e6e92/44361f7b-4609-476f-b4a9-58a7851e6e92.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1007.579462] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27813d6c-ddc8-4c06-9b31-c0750baddf37 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.587070] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1007.587070] env[61985]: value = "task-936315" [ 1007.587070] env[61985]: _type = "Task" [ 1007.587070] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.597492] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936315, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.629028] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1007.629028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.092s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.629028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.520s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.629028] env[61985]: INFO nova.compute.claims [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1007.663156] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936314, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.985257] env[61985]: DEBUG oslo_concurrency.lockutils [req-cfb6b9fd-eaaa-493c-af53-1c91b91dc8ac req-7742fdf5-09b5-426e-afef-7860b68d2ef7 service nova] Releasing lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.985956] env[61985]: DEBUG nova.compute.manager [req-cfb6b9fd-eaaa-493c-af53-1c91b91dc8ac req-7742fdf5-09b5-426e-afef-7860b68d2ef7 service nova] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Received event network-vif-deleted-9f39fdcd-6822-4b1e-b8c8-63d8c3878936 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1008.029407] env[61985]: DEBUG nova.network.neutron [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1008.100930] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936315, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.165772] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936314, 'name': CreateVM_Task, 'duration_secs': 0.529922} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.165943] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1008.167138] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.167399] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.168078] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1008.168419] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be291936-7427-4371-a8cb-96a904e5a7a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.174682] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1008.174682] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f860f1-809e-4126-257b-c222aeca2a13" [ 1008.174682] env[61985]: _type = "Task" [ 1008.174682] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.184437] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f860f1-809e-4126-257b-c222aeca2a13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.297708] env[61985]: DEBUG nova.network.neutron [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Updating instance_info_cache with network_info: [{"id": "63023688-9d9e-4fab-91af-9e9171bccc16", "address": "fa:16:3e:33:8a:78", "network": {"id": "ddf63d11-2633-4be8-b73f-c5ca846b420c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-709607861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8015ef21eaf4a0bb19063e261064cdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63023688-9d", "ovs_interfaceid": "63023688-9d9e-4fab-91af-9e9171bccc16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.598500] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936315, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582579} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.598932] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 44361f7b-4609-476f-b4a9-58a7851e6e92/44361f7b-4609-476f-b4a9-58a7851e6e92.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1008.599228] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1008.599550] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44fab39a-e695-48d0-b80a-9ce70a772e5e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.606416] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1008.606416] env[61985]: value = "task-936316" [ 1008.606416] env[61985]: _type = "Task" [ 1008.606416] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.615594] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936316, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.684632] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f860f1-809e-4126-257b-c222aeca2a13, 'name': SearchDatastore_Task, 'duration_secs': 0.015473} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.685775] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.685775] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.686108] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.686279] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.686514] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.689505] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37700bb6-5d43-4f14-8d95-4825fc9bbff8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.702797] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.703073] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1008.703916] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fbf43a8-2677-41e6-bd9c-e5aa778cbf20 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.712944] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1008.712944] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52064d2c-758a-23c7-1be7-14ca9a8524ae" [ 1008.712944] env[61985]: _type = "Task" [ 1008.712944] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.721503] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52064d2c-758a-23c7-1be7-14ca9a8524ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.800235] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Releasing lock "refresh_cache-90ac7d3e-aef3-4a50-963a-606e7c58f446" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.800469] env[61985]: DEBUG nova.compute.manager [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Instance network_info: |[{"id": "63023688-9d9e-4fab-91af-9e9171bccc16", "address": "fa:16:3e:33:8a:78", "network": {"id": "ddf63d11-2633-4be8-b73f-c5ca846b420c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-709607861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8015ef21eaf4a0bb19063e261064cdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63023688-9d", "ovs_interfaceid": "63023688-9d9e-4fab-91af-9e9171bccc16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1008.800830] env[61985]: DEBUG oslo_concurrency.lockutils [req-2886dc82-f7ce-45ae-bdee-59a22fb75b5d req-5e1c2b53-37a8-4ab1-9c64-8c33f9b85ec1 service nova] Acquired lock "refresh_cache-90ac7d3e-aef3-4a50-963a-606e7c58f446" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.801033] env[61985]: DEBUG nova.network.neutron [req-2886dc82-f7ce-45ae-bdee-59a22fb75b5d req-5e1c2b53-37a8-4ab1-9c64-8c33f9b85ec1 service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Refreshing network info cache for port 63023688-9d9e-4fab-91af-9e9171bccc16 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1008.802264] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:8a:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1470a3f8-be8a-4339-8a6f-9519366f32e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63023688-9d9e-4fab-91af-9e9171bccc16', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.810427] env[61985]: DEBUG oslo.service.loopingcall [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.813910] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1008.814927] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55f9f7f5-839f-4a45-b6d6-de9ad2d17d6d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.840016] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1008.840016] env[61985]: value = "task-936317" [ 1008.840016] env[61985]: _type = "Task" [ 1008.840016] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.852375] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936317, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.985448] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b300c8a7-ce4f-4783-9a3d-5696f89c5984 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.993671] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d2fb6c-5e9c-47fe-bc19-538dd7cb7109 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.027855] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345dc8f0-1048-481c-b5c7-c6ef4006c6e3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.035825] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d506bb-174c-4b17-934d-0a40e41b5f73 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.049678] env[61985]: DEBUG nova.compute.provider_tree [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.118512] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936316, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066639} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.118824] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1009.119623] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f65c81-b214-4832-8e53-bfc6f3e280b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.144401] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 44361f7b-4609-476f-b4a9-58a7851e6e92/44361f7b-4609-476f-b4a9-58a7851e6e92.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1009.144753] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5109f73-4301-42a1-a220-96acf6365281 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.171011] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1009.171011] env[61985]: value = "task-936318" [ 1009.171011] env[61985]: _type = "Task" [ 1009.171011] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.183538] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936318, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.223754] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52064d2c-758a-23c7-1be7-14ca9a8524ae, 'name': SearchDatastore_Task, 'duration_secs': 0.014421} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.224445] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ead8d2ca-2dfd-4eb7-83fd-e87f012bdb44 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.230105] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1009.230105] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5282593c-3bb4-a7c7-2193-24defb2deb77" [ 1009.230105] env[61985]: _type = "Task" [ 1009.230105] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.238211] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5282593c-3bb4-a7c7-2193-24defb2deb77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.352733] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936317, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.582640] env[61985]: DEBUG nova.scheduler.client.report [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 118 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1009.582958] env[61985]: DEBUG nova.compute.provider_tree [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 118 to 119 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1009.583162] env[61985]: DEBUG nova.compute.provider_tree [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.601898] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquiring lock "bed26b5f-7ca5-405a-884e-02b2495dd977" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.602186] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lock "bed26b5f-7ca5-405a-884e-02b2495dd977" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.602423] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquiring lock "bed26b5f-7ca5-405a-884e-02b2495dd977-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.602621] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lock "bed26b5f-7ca5-405a-884e-02b2495dd977-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.602799] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lock "bed26b5f-7ca5-405a-884e-02b2495dd977-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.604980] env[61985]: INFO nova.compute.manager [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Terminating instance [ 1009.606851] env[61985]: DEBUG nova.compute.manager [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1009.607078] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1009.608016] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5d9d7a-36f3-4635-b502-7c13f2d3ef35 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.616310] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1009.616574] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54c6f5d8-d88e-4cb1-9543-d902d61c9d27 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.622580] env[61985]: DEBUG oslo_vmware.api [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for the task: (returnval){ [ 1009.622580] env[61985]: value = "task-936319" [ 1009.622580] env[61985]: _type = "Task" [ 1009.622580] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.633304] env[61985]: DEBUG oslo_vmware.api [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.681415] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936318, 'name': ReconfigVM_Task, 'duration_secs': 0.329137} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.682512] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 44361f7b-4609-476f-b4a9-58a7851e6e92/44361f7b-4609-476f-b4a9-58a7851e6e92.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1009.682512] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3c185f9-1ecd-4b42-a3b7-86daaa83394c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.689790] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1009.689790] env[61985]: value = "task-936320" [ 1009.689790] env[61985]: _type = "Task" [ 1009.689790] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.698646] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936320, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.741445] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5282593c-3bb4-a7c7-2193-24defb2deb77, 'name': SearchDatastore_Task, 'duration_secs': 0.017339} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.741792] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.742137] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 49707589-4969-4e08-882b-2a2c94bc0d85/49707589-4969-4e08-882b-2a2c94bc0d85.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1009.742416] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eda799e4-28ee-4ae7-8a7a-255046d127e9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.749347] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1009.749347] env[61985]: value = "task-936321" [ 1009.749347] env[61985]: _type = "Task" [ 1009.749347] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.758502] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936321, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.759535] env[61985]: DEBUG nova.network.neutron [req-2886dc82-f7ce-45ae-bdee-59a22fb75b5d req-5e1c2b53-37a8-4ab1-9c64-8c33f9b85ec1 service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Updated VIF entry in instance network info cache for port 63023688-9d9e-4fab-91af-9e9171bccc16. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1009.759933] env[61985]: DEBUG nova.network.neutron [req-2886dc82-f7ce-45ae-bdee-59a22fb75b5d req-5e1c2b53-37a8-4ab1-9c64-8c33f9b85ec1 service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Updating instance_info_cache with network_info: [{"id": "63023688-9d9e-4fab-91af-9e9171bccc16", "address": "fa:16:3e:33:8a:78", "network": {"id": "ddf63d11-2633-4be8-b73f-c5ca846b420c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-709607861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8015ef21eaf4a0bb19063e261064cdc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1470a3f8-be8a-4339-8a6f-9519366f32e4", "external-id": "nsx-vlan-transportzone-375", "segmentation_id": 375, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63023688-9d", "ovs_interfaceid": "63023688-9d9e-4fab-91af-9e9171bccc16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.853426] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936317, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.088932] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.089511] env[61985]: DEBUG nova.compute.manager [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1010.092539] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.046s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.092785] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.095064] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 8.959s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.130056] env[61985]: INFO nova.scheduler.client.report [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleted allocations for instance 942e0f09-4d34-4aa2-8a60-4c12f899712c [ 1010.139170] env[61985]: DEBUG oslo_vmware.api [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.201062] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936320, 'name': Rename_Task, 'duration_secs': 0.147068} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.201318] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1010.201567] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53f70dee-49ac-430a-a1df-985084efda62 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.209837] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1010.209837] env[61985]: value = "task-936322" [ 1010.209837] env[61985]: _type = "Task" [ 1010.209837] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.221743] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936322, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.260270] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936321, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.262988] env[61985]: DEBUG oslo_concurrency.lockutils [req-2886dc82-f7ce-45ae-bdee-59a22fb75b5d req-5e1c2b53-37a8-4ab1-9c64-8c33f9b85ec1 service nova] Releasing lock "refresh_cache-90ac7d3e-aef3-4a50-963a-606e7c58f446" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.357164] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936317, 'name': CreateVM_Task, 'duration_secs': 1.484653} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.357552] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1010.358878] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.359282] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.359807] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.360227] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6667251-cc45-4a97-86ee-99ac72cad11f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.367547] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 1010.367547] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b59384-f0ec-d928-ea06-0a124a18fa0f" [ 1010.367547] env[61985]: _type = "Task" [ 1010.367547] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.383209] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b59384-f0ec-d928-ea06-0a124a18fa0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.602944] env[61985]: DEBUG nova.compute.utils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.608409] env[61985]: DEBUG nova.compute.manager [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1010.608409] env[61985]: DEBUG nova.network.neutron [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1010.639375] env[61985]: DEBUG oslo_vmware.api [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936319, 'name': PowerOffVM_Task, 'duration_secs': 0.811783} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.639698] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1010.639894] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1010.643037] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f7e8a00-fd06-49fe-a0bb-8cebf9fe3df8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.648700] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5174fa83-e543-4324-b229-d44e8281c2f9 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "942e0f09-4d34-4aa2-8a60-4c12f899712c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.552s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.707354] env[61985]: DEBUG nova.policy [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a78dbc667d24c819e6447740fc2d8e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2dbceebf07c43ac90ed80f8a09294b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1010.728488] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936322, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.760645] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936321, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.970766} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.760955] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 49707589-4969-4e08-882b-2a2c94bc0d85/49707589-4969-4e08-882b-2a2c94bc0d85.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1010.761185] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1010.761451] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d1c5a8d-5b23-4909-ad10-002ec11d385c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.771297] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1010.771297] env[61985]: value = "task-936324" [ 1010.771297] env[61985]: _type = "Task" [ 1010.771297] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.780113] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936324, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.878965] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b59384-f0ec-d928-ea06-0a124a18fa0f, 'name': SearchDatastore_Task, 'duration_secs': 0.06836} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.879341] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.879581] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.879721] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.879866] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.880055] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.880323] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58b5f8cf-e60e-4630-94d7-93c7bbca0283 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.888789] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.888972] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1010.891999] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ae1fa68-eba6-45cc-a391-3b8511c357db {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.897571] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 1010.897571] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521a0f04-e69a-7763-5efc-1d556e332e06" [ 1010.897571] env[61985]: _type = "Task" [ 1010.897571] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.905797] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521a0f04-e69a-7763-5efc-1d556e332e06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.943925] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13dfcd8e-5c92-4dcb-be5c-52374b70407a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.952170] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36ffad3-c467-4b31-8fa8-924c27401226 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.983302] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4313393-d82c-4a95-b39d-35e8a9811c37 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.991339] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2ba60f-49af-4076-9af7-c63d70691c7c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.006087] env[61985]: DEBUG nova.compute.provider_tree [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.052585] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "986c576e-be02-48ac-b24c-72edccab25be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.052925] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "986c576e-be02-48ac-b24c-72edccab25be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.053165] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "986c576e-be02-48ac-b24c-72edccab25be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.053392] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "986c576e-be02-48ac-b24c-72edccab25be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.053561] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "986c576e-be02-48ac-b24c-72edccab25be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.056126] env[61985]: INFO nova.compute.manager [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Terminating instance [ 1011.059142] env[61985]: DEBUG nova.compute.manager [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1011.059347] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1011.060222] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6c7407-eda2-4c1b-ab26-4cd46c05cc51 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.069392] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1011.069957] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-076132e3-0203-4484-9669-22872816df5d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.076281] env[61985]: DEBUG oslo_vmware.api [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1011.076281] env[61985]: value = "task-936325" [ 1011.076281] env[61985]: _type = "Task" [ 1011.076281] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.084773] env[61985]: DEBUG oslo_vmware.api [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.107778] env[61985]: DEBUG nova.compute.manager [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1011.220762] env[61985]: DEBUG oslo_vmware.api [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936322, 'name': PowerOnVM_Task, 'duration_secs': 0.819446} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.221055] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1011.222807] env[61985]: INFO nova.compute.manager [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Took 7.73 seconds to spawn the instance on the hypervisor. [ 1011.222807] env[61985]: DEBUG nova.compute.manager [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1011.222807] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e759a5ef-5660-43a8-8c50-032b8a65bf91 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.287464] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936324, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.236095} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.287819] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1011.288984] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15010727-3a50-4fe1-97e4-8b02fc59ad08 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.316655] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 49707589-4969-4e08-882b-2a2c94bc0d85/49707589-4969-4e08-882b-2a2c94bc0d85.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.317573] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-defdc24c-c590-426e-9ff9-755e435a61fb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.333477] env[61985]: DEBUG nova.network.neutron [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Successfully created port: 88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1011.342426] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1011.342426] env[61985]: value = "task-936326" [ 1011.342426] env[61985]: _type = "Task" [ 1011.342426] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.353078] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936326, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.407794] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521a0f04-e69a-7763-5efc-1d556e332e06, 'name': SearchDatastore_Task, 'duration_secs': 0.010385} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.408651] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2fa3a77-b630-4a98-9b67-8923ec384958 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.413862] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 1011.413862] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52702e08-827d-e2b6-df24-4fdd3d95f94a" [ 1011.413862] env[61985]: _type = "Task" [ 1011.413862] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.421663] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52702e08-827d-e2b6-df24-4fdd3d95f94a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.509445] env[61985]: DEBUG nova.scheduler.client.report [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.586529] env[61985]: DEBUG oslo_vmware.api [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.756406] env[61985]: INFO nova.compute.manager [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Took 34.87 seconds to build instance. [ 1011.854225] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936326, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.867031] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.869031] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.748171] env[61985]: DEBUG nova.compute.manager [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1012.748171] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c859b6c7-1713-421c-9494-fd1318e588a4 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.871s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.748547] env[61985]: DEBUG nova.compute.manager [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1012.756222] env[61985]: DEBUG nova.compute.manager [req-896bd347-6800-41d2-a93c-190187a07a61 req-6bafb9fa-2bb5-4a60-89d1-6cc80fd96fa2 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Received event network-changed-e35f2970-0b73-4b31-925f-56ae8e0e7f39 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1012.756604] env[61985]: DEBUG nova.compute.manager [req-896bd347-6800-41d2-a93c-190187a07a61 req-6bafb9fa-2bb5-4a60-89d1-6cc80fd96fa2 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Refreshing instance network info cache due to event network-changed-e35f2970-0b73-4b31-925f-56ae8e0e7f39. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1012.758028] env[61985]: DEBUG oslo_concurrency.lockutils [req-896bd347-6800-41d2-a93c-190187a07a61 req-6bafb9fa-2bb5-4a60-89d1-6cc80fd96fa2 service nova] Acquiring lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.758028] env[61985]: DEBUG oslo_concurrency.lockutils [req-896bd347-6800-41d2-a93c-190187a07a61 req-6bafb9fa-2bb5-4a60-89d1-6cc80fd96fa2 service nova] Acquired lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.758028] env[61985]: DEBUG nova.network.neutron [req-896bd347-6800-41d2-a93c-190187a07a61 req-6bafb9fa-2bb5-4a60-89d1-6cc80fd96fa2 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Refreshing network info cache for port e35f2970-0b73-4b31-925f-56ae8e0e7f39 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1012.775234] env[61985]: DEBUG oslo_vmware.api [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936325, 'name': PowerOffVM_Task, 'duration_secs': 1.115931} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.782206] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1012.782419] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1012.783091] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936326, 'name': ReconfigVM_Task, 'duration_secs': 0.986698} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.783351] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52702e08-827d-e2b6-df24-4fdd3d95f94a, 'name': SearchDatastore_Task, 'duration_secs': 0.013736} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.783895] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a69a7297-983d-4f9f-9085-c5616d544c32 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.785671] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 49707589-4969-4e08-882b-2a2c94bc0d85/49707589-4969-4e08-882b-2a2c94bc0d85.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.786365] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.786633] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 90ac7d3e-aef3-4a50-963a-606e7c58f446/90ac7d3e-aef3-4a50-963a-606e7c58f446.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1012.787151] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-32c3aa12-58fc-46ef-8399-04ea48c5e4e8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.788920] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f6bb0a8-b6b3-47b7-b979-812cae9728eb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.800343] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 1012.800343] env[61985]: value = "task-936329" [ 1012.800343] env[61985]: _type = "Task" [ 1012.800343] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.803172] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.803455] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.803666] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.803951] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.804165] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.804378] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.804652] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.804879] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.805122] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.805353] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.805611] env[61985]: DEBUG nova.virt.hardware [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.809171] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b864866-6f2a-4d28-b2d3-40f28b9e8f68 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.812421] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1012.812421] env[61985]: value = "task-936328" [ 1012.812421] env[61985]: _type = "Task" [ 1012.812421] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.826990] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ae4351-e3fc-4f5d-bdff-a5fcfc329b93 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.831513] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936329, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.835037] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936328, 'name': Rename_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.262594] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 3.166s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.268536] env[61985]: DEBUG oslo_concurrency.lockutils [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.490s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.269090] env[61985]: DEBUG oslo_concurrency.lockutils [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.271465] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.765s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.271918] env[61985]: DEBUG nova.objects.instance [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lazy-loading 'resources' on Instance uuid a77ca23f-b2c0-4822-8e48-3e47e0dadb27 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.295656] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.310178] env[61985]: INFO nova.scheduler.client.report [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Deleted allocations for instance 6dd8218b-c14d-40c9-87df-097fab06c669 [ 1013.326705] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936329, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.332537] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936328, 'name': Rename_Task, 'duration_secs': 0.217332} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.333213] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1013.334124] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0904686e-fe66-479b-b7ad-3bb31df9c5ea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.341388] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1013.341388] env[61985]: value = "task-936330" [ 1013.341388] env[61985]: _type = "Task" [ 1013.341388] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.356456] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936330, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.823344] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936329, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.613741} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.824678] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 90ac7d3e-aef3-4a50-963a-606e7c58f446/90ac7d3e-aef3-4a50-963a-606e7c58f446.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1013.824871] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.825458] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-374aae07-46d6-4c5f-a3a6-9b7adb7fdcf7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.833486] env[61985]: INFO nova.scheduler.client.report [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted allocation for migration 1bf48126-5cd8-40dd-bbe0-d7c7e1d54f35 [ 1013.843631] env[61985]: DEBUG oslo_concurrency.lockutils [None req-11a9a455-9039-416e-9cff-95d70c07d12b tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "6dd8218b-c14d-40c9-87df-097fab06c669" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.993s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.852659] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 1013.852659] env[61985]: value = "task-936331" [ 1013.852659] env[61985]: _type = "Task" [ 1013.852659] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.863050] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936330, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.868153] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936331, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.000644] env[61985]: DEBUG nova.network.neutron [req-896bd347-6800-41d2-a93c-190187a07a61 req-6bafb9fa-2bb5-4a60-89d1-6cc80fd96fa2 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updated VIF entry in instance network info cache for port e35f2970-0b73-4b31-925f-56ae8e0e7f39. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1014.001084] env[61985]: DEBUG nova.network.neutron [req-896bd347-6800-41d2-a93c-190187a07a61 req-6bafb9fa-2bb5-4a60-89d1-6cc80fd96fa2 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance_info_cache with network_info: [{"id": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "address": "fa:16:3e:13:20:58", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35f2970-0b", "ovs_interfaceid": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.025245] env[61985]: DEBUG nova.compute.manager [req-575805e4-f614-4578-99d8-39ed6adadb04 req-5142c4e8-371a-47e1-ade8-2db57aba332d service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Received event network-vif-plugged-88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1014.025245] env[61985]: DEBUG oslo_concurrency.lockutils [req-575805e4-f614-4578-99d8-39ed6adadb04 req-5142c4e8-371a-47e1-ade8-2db57aba332d service nova] Acquiring lock "8db82979-e527-4c63-8ce8-fe1456f7708c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.025245] env[61985]: DEBUG oslo_concurrency.lockutils [req-575805e4-f614-4578-99d8-39ed6adadb04 req-5142c4e8-371a-47e1-ade8-2db57aba332d service nova] Lock "8db82979-e527-4c63-8ce8-fe1456f7708c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.025245] env[61985]: DEBUG oslo_concurrency.lockutils [req-575805e4-f614-4578-99d8-39ed6adadb04 req-5142c4e8-371a-47e1-ade8-2db57aba332d service nova] Lock "8db82979-e527-4c63-8ce8-fe1456f7708c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.025245] env[61985]: DEBUG nova.compute.manager [req-575805e4-f614-4578-99d8-39ed6adadb04 req-5142c4e8-371a-47e1-ade8-2db57aba332d service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] No waiting events found dispatching network-vif-plugged-88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.025245] env[61985]: WARNING nova.compute.manager [req-575805e4-f614-4578-99d8-39ed6adadb04 req-5142c4e8-371a-47e1-ade8-2db57aba332d service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Received unexpected event network-vif-plugged-88a63ff6-55f6-4b8b-ba9e-188245f7ea87 for instance with vm_state building and task_state spawning. [ 1014.031945] env[61985]: DEBUG nova.network.neutron [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Successfully updated port: 88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1014.128197] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73142a80-55ad-46e0-b43c-c5dc92e76147 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.135749] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2afe78-b446-4d2d-891a-a62e151f6624 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.171213] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2db28f9-82b9-4851-a19d-a1c9fc2e19d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.178987] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08354b7d-bef1-4d89-a886-f96dee47e7ea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.193979] env[61985]: DEBUG nova.compute.provider_tree [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.348016] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fd67bc79-72d2-4cb7-b95a-85415e7ecb54 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 16.171s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.359101] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936330, 'name': PowerOnVM_Task, 'duration_secs': 0.632487} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.362668] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1014.362900] env[61985]: DEBUG nova.compute.manager [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1014.363910] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429e7a9d-c1b1-4bf2-85c6-e1ecbc2d641a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.371127] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936331, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076052} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.372913] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.376454] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10af0308-dc1b-4701-9980-998695bc9a7e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.400544] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 90ac7d3e-aef3-4a50-963a-606e7c58f446/90ac7d3e-aef3-4a50-963a-606e7c58f446.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.400885] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3afd8a62-7196-439b-9a2a-0f209b18e4ef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.422654] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 1014.422654] env[61985]: value = "task-936332" [ 1014.422654] env[61985]: _type = "Task" [ 1014.422654] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.431501] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936332, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.507216] env[61985]: DEBUG oslo_concurrency.lockutils [req-896bd347-6800-41d2-a93c-190187a07a61 req-6bafb9fa-2bb5-4a60-89d1-6cc80fd96fa2 service nova] Releasing lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.534808] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.534975] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.535160] env[61985]: DEBUG nova.network.neutron [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1014.700763] env[61985]: DEBUG nova.scheduler.client.report [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.729964] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1014.730267] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1014.730474] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Deleting the datastore file [datastore1] bed26b5f-7ca5-405a-884e-02b2495dd977 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.730797] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82c91877-b9ab-4ad6-a35a-26b03466731d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.742030] env[61985]: DEBUG oslo_vmware.api [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for the task: (returnval){ [ 1014.742030] env[61985]: value = "task-936333" [ 1014.742030] env[61985]: _type = "Task" [ 1014.742030] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.745556] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1014.745556] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1014.745556] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleting the datastore file [datastore1] 986c576e-be02-48ac-b24c-72edccab25be {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.745556] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-23dff7b9-a3bd-49ff-babd-ac112165eb05 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.757154] env[61985]: DEBUG oslo_vmware.api [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936333, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.759935] env[61985]: DEBUG oslo_vmware.api [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1014.759935] env[61985]: value = "task-936334" [ 1014.759935] env[61985]: _type = "Task" [ 1014.759935] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.771521] env[61985]: DEBUG oslo_vmware.api [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936334, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.882284] env[61985]: INFO nova.compute.manager [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] bringing vm to original state: 'stopped' [ 1014.914707] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525fe5ee-a9ce-6e08-8184-5b3421e34878/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1014.924019] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75b55f0-0e60-4377-9eaa-02925ecdbebb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.933021] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525fe5ee-a9ce-6e08-8184-5b3421e34878/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1014.933021] env[61985]: ERROR oslo_vmware.rw_handles [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525fe5ee-a9ce-6e08-8184-5b3421e34878/disk-0.vmdk due to incomplete transfer. [ 1014.935781] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c08d8066-ff90-4c62-8866-e6e926480eb3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.937811] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936332, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.947020] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525fe5ee-a9ce-6e08-8184-5b3421e34878/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1014.947020] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Uploaded image 10a07ac1-b198-4338-8b14-ed1bbddb0250 to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1014.949907] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1014.950936] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c1cc0f94-25ba-43c5-831b-be0bc1a302c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.957726] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1014.957726] env[61985]: value = "task-936336" [ 1014.957726] env[61985]: _type = "Task" [ 1014.957726] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.969469] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936336, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.091036] env[61985]: DEBUG nova.network.neutron [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1015.207074] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.935s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.210147] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.390s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.210615] env[61985]: DEBUG nova.objects.instance [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lazy-loading 'resources' on Instance uuid b6eb50ff-e685-4e8a-92f5-c2661136c361 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.231407] env[61985]: INFO nova.scheduler.client.report [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Deleted allocations for instance a77ca23f-b2c0-4822-8e48-3e47e0dadb27 [ 1015.253791] env[61985]: DEBUG oslo_vmware.api [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Task: {'id': task-936333, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.285749} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.254591] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1015.254834] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1015.255179] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1015.255286] env[61985]: INFO nova.compute.manager [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Took 5.65 seconds to destroy the instance on the hypervisor. [ 1015.255467] env[61985]: DEBUG oslo.service.loopingcall [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.255691] env[61985]: DEBUG nova.compute.manager [-] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1015.255789] env[61985]: DEBUG nova.network.neutron [-] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1015.272156] env[61985]: DEBUG oslo_vmware.api [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936334, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.28989} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.272425] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1015.272846] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1015.273102] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1015.273292] env[61985]: INFO nova.compute.manager [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Took 4.21 seconds to destroy the instance on the hypervisor. [ 1015.273540] env[61985]: DEBUG oslo.service.loopingcall [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.274099] env[61985]: DEBUG nova.compute.manager [-] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1015.274727] env[61985]: DEBUG nova.network.neutron [-] [instance: 986c576e-be02-48ac-b24c-72edccab25be] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1015.435964] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936332, 'name': ReconfigVM_Task, 'duration_secs': 0.571239} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.435964] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 90ac7d3e-aef3-4a50-963a-606e7c58f446/90ac7d3e-aef3-4a50-963a-606e7c58f446.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.435964] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6678a856-0dd8-47a8-8866-5534da508ced {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.442330] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 1015.442330] env[61985]: value = "task-936337" [ 1015.442330] env[61985]: _type = "Task" [ 1015.442330] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.455615] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936337, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.469711] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936336, 'name': Destroy_Task} progress is 33%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.491041] env[61985]: DEBUG nova.network.neutron [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updating instance_info_cache with network_info: [{"id": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "address": "fa:16:3e:16:ec:4d", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88a63ff6-55", "ovs_interfaceid": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.740887] env[61985]: DEBUG oslo_concurrency.lockutils [None req-89e3b8d9-1b97-438d-a16b-8fb8f39976a8 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "a77ca23f-b2c0-4822-8e48-3e47e0dadb27" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.567s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.835419] env[61985]: DEBUG nova.compute.manager [req-7ef87be1-9f11-46b4-b5fb-49f44d35c85c req-5206a8b5-ead3-4188-91d5-4a33014cc054 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Received event network-vif-deleted-334af3b0-2de9-4794-890e-ea3f7b0dd351 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1015.835803] env[61985]: INFO nova.compute.manager [req-7ef87be1-9f11-46b4-b5fb-49f44d35c85c req-5206a8b5-ead3-4188-91d5-4a33014cc054 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Neutron deleted interface 334af3b0-2de9-4794-890e-ea3f7b0dd351; detaching it from the instance and deleting it from the info cache [ 1015.836125] env[61985]: DEBUG nova.network.neutron [req-7ef87be1-9f11-46b4-b5fb-49f44d35c85c req-5206a8b5-ead3-4188-91d5-4a33014cc054 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.890909] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "49707589-4969-4e08-882b-2a2c94bc0d85" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.891215] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "49707589-4969-4e08-882b-2a2c94bc0d85" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.891413] env[61985]: DEBUG nova.compute.manager [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.892380] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1cfe9a0-cb5a-4ec4-b974-6b041a5efda1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.902139] env[61985]: DEBUG nova.compute.manager [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61985) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3374}} [ 1015.905131] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1015.905408] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0691de04-91e3-4b3c-9a11-ce773c17db84 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.913560] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1015.913560] env[61985]: value = "task-936338" [ 1015.913560] env[61985]: _type = "Task" [ 1015.913560] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.926106] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936338, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.953115] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936337, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.968803] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936336, 'name': Destroy_Task, 'duration_secs': 0.791563} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.969169] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Destroyed the VM [ 1015.969510] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1015.969796] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ec6f0a14-20f0-4aa8-a152-b9844d016345 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.977218] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1015.977218] env[61985]: value = "task-936339" [ 1015.977218] env[61985]: _type = "Task" [ 1015.977218] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.990113] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936339, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.996249] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.996558] env[61985]: DEBUG nova.compute.manager [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Instance network_info: |[{"id": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "address": "fa:16:3e:16:ec:4d", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88a63ff6-55", "ovs_interfaceid": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1015.997265] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:ec:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88a63ff6-55f6-4b8b-ba9e-188245f7ea87', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1016.006828] env[61985]: DEBUG oslo.service.loopingcall [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.006828] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1016.006828] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13cd3878-633f-47bd-9207-30b01b0230eb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.029843] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1016.029843] env[61985]: value = "task-936340" [ 1016.029843] env[61985]: _type = "Task" [ 1016.029843] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.045381] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936340, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.083245] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe20678-d71a-4b01-908a-aeb95c784c59 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.090301] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d7e707-49cd-4aec-be36-a2629fb7dd6a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.126808] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee42771-8722-4a2b-8cba-1dfc8cc22fcb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.141364] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc6d1d01-5f21-4b7e-8d56-51a95fa9a933 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.150116] env[61985]: DEBUG nova.network.neutron [-] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.167471] env[61985]: DEBUG nova.compute.provider_tree [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.182771] env[61985]: DEBUG nova.compute.manager [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Received event network-changed-88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1016.182996] env[61985]: DEBUG nova.compute.manager [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Refreshing instance network info cache due to event network-changed-88a63ff6-55f6-4b8b-ba9e-188245f7ea87. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1016.183313] env[61985]: DEBUG oslo_concurrency.lockutils [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] Acquiring lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.183536] env[61985]: DEBUG oslo_concurrency.lockutils [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] Acquired lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.183712] env[61985]: DEBUG nova.network.neutron [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Refreshing network info cache for port 88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1016.306354] env[61985]: DEBUG nova.network.neutron [-] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.339782] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6beea71a-68d1-4e35-90b1-33111c86d821 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.348988] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0170d6ad-d038-40d5-897e-7db0cf5d5b04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.393908] env[61985]: DEBUG nova.compute.manager [req-7ef87be1-9f11-46b4-b5fb-49f44d35c85c req-5206a8b5-ead3-4188-91d5-4a33014cc054 service nova] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Detach interface failed, port_id=334af3b0-2de9-4794-890e-ea3f7b0dd351, reason: Instance 986c576e-be02-48ac-b24c-72edccab25be could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1016.430841] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936338, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.453927] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936337, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.486909] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936339, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.540382] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936340, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.636712] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "625a3143-d138-4b52-aeb7-9e365c6f1128" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.637052] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.637302] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "625a3143-d138-4b52-aeb7-9e365c6f1128-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.637508] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.637731] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.640155] env[61985]: INFO nova.compute.manager [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Terminating instance [ 1016.642679] env[61985]: DEBUG nova.compute.manager [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1016.642912] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1016.643821] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f95436a-33ba-4669-83ce-b9ebf20708ba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.652087] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1016.652388] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2d08fee-8b7c-441d-b166-11dcfc947e1f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.660915] env[61985]: DEBUG oslo_vmware.api [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1016.660915] env[61985]: value = "task-936341" [ 1016.660915] env[61985]: _type = "Task" [ 1016.660915] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.673358] env[61985]: INFO nova.compute.manager [-] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Took 1.40 seconds to deallocate network for instance. [ 1016.673358] env[61985]: DEBUG nova.scheduler.client.report [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.678069] env[61985]: DEBUG oslo_vmware.api [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936341, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.797064] env[61985]: DEBUG oslo_concurrency.lockutils [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "fea0cf39-e851-409f-86f5-31cc128a44dc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.797739] env[61985]: DEBUG oslo_concurrency.lockutils [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.797951] env[61985]: DEBUG oslo_concurrency.lockutils [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "fea0cf39-e851-409f-86f5-31cc128a44dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.798344] env[61985]: DEBUG oslo_concurrency.lockutils [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.798692] env[61985]: DEBUG oslo_concurrency.lockutils [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.802617] env[61985]: INFO nova.compute.manager [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Terminating instance [ 1016.806073] env[61985]: DEBUG nova.compute.manager [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1016.806736] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1016.808160] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd47b3d-e448-4a26-a9a3-f2ab3711ecf3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.815285] env[61985]: INFO nova.compute.manager [-] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Took 1.56 seconds to deallocate network for instance. [ 1016.827786] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1016.828339] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc0b7f85-16e7-4667-a890-7dbc9c738875 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.838034] env[61985]: DEBUG oslo_vmware.api [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 1016.838034] env[61985]: value = "task-936342" [ 1016.838034] env[61985]: _type = "Task" [ 1016.838034] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.851034] env[61985]: DEBUG oslo_vmware.api [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936342, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.927185] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936338, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.956284] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936337, 'name': Rename_Task, 'duration_secs': 1.322534} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.956609] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1016.956875] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbb1b2d5-619f-4c84-b99c-c7b12266c87f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.964238] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 1016.964238] env[61985]: value = "task-936343" [ 1016.964238] env[61985]: _type = "Task" [ 1016.964238] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.976343] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936343, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.993876] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936339, 'name': RemoveSnapshot_Task, 'duration_secs': 0.821763} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.993876] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1016.993876] env[61985]: DEBUG nova.compute.manager [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.993876] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4fdff88-9766-4f48-82ca-7ece72ba5a55 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.039491] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936340, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.172458] env[61985]: DEBUG oslo_vmware.api [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936341, 'name': PowerOffVM_Task, 'duration_secs': 0.209339} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.172814] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1017.173019] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1017.173344] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9fe4fe3-442a-4df8-9c23-a13ae6b2213d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.175775] env[61985]: DEBUG nova.network.neutron [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updated VIF entry in instance network info cache for port 88a63ff6-55f6-4b8b-ba9e-188245f7ea87. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.176133] env[61985]: DEBUG nova.network.neutron [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updating instance_info_cache with network_info: [{"id": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "address": "fa:16:3e:16:ec:4d", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88a63ff6-55", "ovs_interfaceid": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.181204] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.968s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.181204] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.956s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.181924] env[61985]: INFO nova.compute.claims [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.184976] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.199189] env[61985]: INFO nova.scheduler.client.report [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted allocations for instance b6eb50ff-e685-4e8a-92f5-c2661136c361 [ 1017.250863] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1017.250863] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1017.250863] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleting the datastore file [datastore1] 625a3143-d138-4b52-aeb7-9e365c6f1128 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1017.250863] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53baf7ed-0212-433a-8de0-4e00ff75b952 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.257096] env[61985]: DEBUG oslo_vmware.api [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1017.257096] env[61985]: value = "task-936345" [ 1017.257096] env[61985]: _type = "Task" [ 1017.257096] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.265781] env[61985]: DEBUG oslo_vmware.api [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936345, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.331135] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.351682] env[61985]: DEBUG oslo_vmware.api [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936342, 'name': PowerOffVM_Task, 'duration_secs': 0.209419} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.351981] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1017.352159] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1017.352419] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5f64c0d-aabb-4630-802e-603aa28e0075 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.365137] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "8c4a1fe4-4ded-4064-8261-24826181a68c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.365376] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.427470] env[61985]: DEBUG oslo_vmware.api [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936338, 'name': PowerOffVM_Task, 'duration_secs': 1.02693} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.427834] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1017.428034] env[61985]: DEBUG nova.compute.manager [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.428817] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073d31e9-7e74-4bfb-9128-3cf66f6317df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.477732] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936343, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.508405] env[61985]: INFO nova.compute.manager [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Shelve offloading [ 1017.510125] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1017.510371] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f91c58b4-b6be-40ed-9696-784e9d22d298 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.517745] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1017.517745] env[61985]: value = "task-936347" [ 1017.517745] env[61985]: _type = "Task" [ 1017.517745] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.525091] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936347, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.538868] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936340, 'name': CreateVM_Task, 'duration_secs': 1.046232} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.539045] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1017.539678] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.539861] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.540213] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.540743] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66a22d42-d0ba-41c6-8cde-c777b24d882d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.544928] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1017.544928] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524d598a-eaf8-e1ca-8acd-fc4d066e48a1" [ 1017.544928] env[61985]: _type = "Task" [ 1017.544928] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.552315] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524d598a-eaf8-e1ca-8acd-fc4d066e48a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.681691] env[61985]: DEBUG oslo_concurrency.lockutils [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] Releasing lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.682145] env[61985]: DEBUG nova.compute.manager [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Received event network-vif-deleted-c38204b9-b8fd-45ce-8fb5-f1b64f8e1649 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1017.682340] env[61985]: INFO nova.compute.manager [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Neutron deleted interface c38204b9-b8fd-45ce-8fb5-f1b64f8e1649; detaching it from the instance and deleting it from the info cache [ 1017.682534] env[61985]: DEBUG nova.network.neutron [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.706196] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8f37b4f1-ddb3-497d-aba6-b5b00aec13f7 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "b6eb50ff-e685-4e8a-92f5-c2661136c361" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.959s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.764817] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1017.765276] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1017.765538] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Deleting the datastore file [datastore2] fea0cf39-e851-409f-86f5-31cc128a44dc {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1017.766305] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a66d57e9-17b8-4783-8423-078b8db3c066 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.772067] env[61985]: DEBUG oslo_vmware.api [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936345, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.32151} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.772635] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1017.772832] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1017.773048] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1017.773234] env[61985]: INFO nova.compute.manager [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1017.773475] env[61985]: DEBUG oslo.service.loopingcall [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.773671] env[61985]: DEBUG nova.compute.manager [-] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1017.773767] env[61985]: DEBUG nova.network.neutron [-] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1017.776508] env[61985]: DEBUG oslo_vmware.api [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for the task: (returnval){ [ 1017.776508] env[61985]: value = "task-936348" [ 1017.776508] env[61985]: _type = "Task" [ 1017.776508] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.785669] env[61985]: DEBUG oslo_vmware.api [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936348, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.868312] env[61985]: DEBUG nova.compute.manager [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1017.943102] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "49707589-4969-4e08-882b-2a2c94bc0d85" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.052s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.978141] env[61985]: DEBUG oslo_vmware.api [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936343, 'name': PowerOnVM_Task, 'duration_secs': 0.711367} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.978488] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1017.978603] env[61985]: INFO nova.compute.manager [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Took 12.40 seconds to spawn the instance on the hypervisor. [ 1017.978784] env[61985]: DEBUG nova.compute.manager [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.979581] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba28dad-fae5-417a-86de-9c597bfa8bfc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.036567] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1018.036783] env[61985]: DEBUG nova.compute.manager [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.037565] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ec72087-833e-41c0-8413-f734805dd939 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.044505] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.044730] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.044835] env[61985]: DEBUG nova.network.neutron [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1018.055969] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524d598a-eaf8-e1ca-8acd-fc4d066e48a1, 'name': SearchDatastore_Task, 'duration_secs': 0.012537} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.056297] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.056545] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.057028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.057028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.057151] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.057670] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd01e455-885c-4322-a291-22330934104e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.067034] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.067336] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1018.070140] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f3e0844-d362-4a24-a196-efe9f32e9cdc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.075963] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1018.075963] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b59ef4-34ab-9312-1aec-f1ccb6c3587c" [ 1018.075963] env[61985]: _type = "Task" [ 1018.075963] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.088077] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b59ef4-34ab-9312-1aec-f1ccb6c3587c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.186818] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-185815dd-ac5b-4789-818a-6dbe1e8429c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.198994] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a815a3e-f212-445f-a8fc-4795f78cb925 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.229548] env[61985]: DEBUG nova.compute.manager [req-42c08223-4b1d-45bb-9225-7e4f42622792 req-26b08570-1f54-4907-85e1-77faa25c8c2f service nova] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Detach interface failed, port_id=c38204b9-b8fd-45ce-8fb5-f1b64f8e1649, reason: Instance bed26b5f-7ca5-405a-884e-02b2495dd977 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1018.285718] env[61985]: DEBUG oslo_vmware.api [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Task: {'id': task-936348, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.492955} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.288075] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1018.288281] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1018.288467] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1018.288647] env[61985]: INFO nova.compute.manager [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Took 1.48 seconds to destroy the instance on the hypervisor. [ 1018.288893] env[61985]: DEBUG oslo.service.loopingcall [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.289530] env[61985]: DEBUG nova.compute.manager [-] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1018.289628] env[61985]: DEBUG nova.network.neutron [-] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1018.393319] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.453416] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.483997] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ffa9d9-dad1-44fb-9bc3-e0d1259e8ab5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.495934] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80be1cd5-e56c-49bc-a742-5335d970d421 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.503237] env[61985]: INFO nova.compute.manager [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Took 36.86 seconds to build instance. [ 1018.540095] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ff8d03-d6d5-443d-98fc-e2d09b89e1b2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.550119] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a96e3a2-d115-4d16-97d2-12f9de370072 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.569218] env[61985]: DEBUG nova.compute.provider_tree [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.586641] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b59ef4-34ab-9312-1aec-f1ccb6c3587c, 'name': SearchDatastore_Task, 'duration_secs': 0.011944} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.589466] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d274690-b81a-4da2-9135-179a00e01bcb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.595132] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1018.595132] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e849f4-8100-2b4b-f7c2-4a3eeb386893" [ 1018.595132] env[61985]: _type = "Task" [ 1018.595132] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.603152] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e849f4-8100-2b4b-f7c2-4a3eeb386893, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.783037] env[61985]: DEBUG nova.network.neutron [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Updating instance_info_cache with network_info: [{"id": "8e692ff4-8a38-44ec-ac4b-05e0bda78308", "address": "fa:16:3e:80:98:79", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e692ff4-8a", "ovs_interfaceid": "8e692ff4-8a38-44ec-ac4b-05e0bda78308", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.870500] env[61985]: DEBUG nova.compute.manager [req-c2d91be3-09e2-47c6-a41b-9b0aef77e261 req-e5941304-3cb9-432b-b83f-45167c119e61 service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Received event network-vif-deleted-262813ed-9baf-4240-b57d-fc4a648b5532 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1018.870960] env[61985]: INFO nova.compute.manager [req-c2d91be3-09e2-47c6-a41b-9b0aef77e261 req-e5941304-3cb9-432b-b83f-45167c119e61 service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Neutron deleted interface 262813ed-9baf-4240-b57d-fc4a648b5532; detaching it from the instance and deleting it from the info cache [ 1018.871198] env[61985]: DEBUG nova.network.neutron [req-c2d91be3-09e2-47c6-a41b-9b0aef77e261 req-e5941304-3cb9-432b-b83f-45167c119e61 service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.891292] env[61985]: DEBUG nova.compute.manager [req-79bd6f97-5b30-4d7a-bc6b-f2d5891b0f85 req-ba0e3508-e729-4812-93a6-6fe47e79d916 service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Received event network-vif-deleted-adde3d48-ebd4-450d-920a-fb93eeff4a8b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1018.891572] env[61985]: INFO nova.compute.manager [req-79bd6f97-5b30-4d7a-bc6b-f2d5891b0f85 req-ba0e3508-e729-4812-93a6-6fe47e79d916 service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Neutron deleted interface adde3d48-ebd4-450d-920a-fb93eeff4a8b; detaching it from the instance and deleting it from the info cache [ 1018.892098] env[61985]: DEBUG nova.network.neutron [req-79bd6f97-5b30-4d7a-bc6b-f2d5891b0f85 req-ba0e3508-e729-4812-93a6-6fe47e79d916 service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.005571] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7969e8ec-f89e-497c-99ec-851e536bcf58 tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "90ac7d3e-aef3-4a50-963a-606e7c58f446" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.367s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.011027] env[61985]: DEBUG nova.network.neutron [-] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.055237] env[61985]: DEBUG nova.network.neutron [-] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.091390] env[61985]: ERROR nova.scheduler.client.report [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [req-043e992c-3949-4b31-9fea-27b1f91a3482] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID aed7e5b3-c662-4538-8447-c4f67b460215. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-043e992c-3949-4b31-9fea-27b1f91a3482"}]} [ 1019.106225] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e849f4-8100-2b4b-f7c2-4a3eeb386893, 'name': SearchDatastore_Task, 'duration_secs': 0.016185} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.106521] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.106793] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8db82979-e527-4c63-8ce8-fe1456f7708c/8db82979-e527-4c63-8ce8-fe1456f7708c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1019.107076] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f48d46e7-21c4-4bb1-ab4a-8584dd6d9884 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.112311] env[61985]: DEBUG nova.scheduler.client.report [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1019.118219] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1019.118219] env[61985]: value = "task-936349" [ 1019.118219] env[61985]: _type = "Task" [ 1019.118219] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.126468] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936349, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.127471] env[61985]: DEBUG nova.scheduler.client.report [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1019.127775] env[61985]: DEBUG nova.compute.provider_tree [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 42, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1019.139878] env[61985]: DEBUG nova.scheduler.client.report [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1019.164932] env[61985]: DEBUG nova.scheduler.client.report [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1019.228393] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.228656] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.286373] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.374751] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67ef3947-0102-4b8a-8299-a4a1aa51e12d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.386582] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ec1112-6abd-491a-bc76-529302e44678 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.405135] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f57ea546-d647-4be1-82e4-9281e111e680 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.420022] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16bd8a94-cd43-4135-8ef2-0d4c53257868 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.451463] env[61985]: DEBUG nova.compute.manager [req-c2d91be3-09e2-47c6-a41b-9b0aef77e261 req-e5941304-3cb9-432b-b83f-45167c119e61 service nova] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Detach interface failed, port_id=262813ed-9baf-4240-b57d-fc4a648b5532, reason: Instance 625a3143-d138-4b52-aeb7-9e365c6f1128 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1019.472290] env[61985]: DEBUG nova.compute.manager [req-79bd6f97-5b30-4d7a-bc6b-f2d5891b0f85 req-ba0e3508-e729-4812-93a6-6fe47e79d916 service nova] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Detach interface failed, port_id=adde3d48-ebd4-450d-920a-fb93eeff4a8b, reason: Instance fea0cf39-e851-409f-86f5-31cc128a44dc could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1019.515669] env[61985]: INFO nova.compute.manager [-] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Took 1.74 seconds to deallocate network for instance. [ 1019.529181] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b41f12e-e420-4584-9802-2ed84275d545 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.535934] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ef9464-2d99-47b7-aa44-0986a36a22f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.571673] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "90ac7d3e-aef3-4a50-963a-606e7c58f446" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.571947] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "90ac7d3e-aef3-4a50-963a-606e7c58f446" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.572226] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "90ac7d3e-aef3-4a50-963a-606e7c58f446-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.572433] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "90ac7d3e-aef3-4a50-963a-606e7c58f446-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.572611] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "90ac7d3e-aef3-4a50-963a-606e7c58f446-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.576803] env[61985]: INFO nova.compute.manager [-] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Took 1.28 seconds to deallocate network for instance. [ 1019.576803] env[61985]: INFO nova.compute.manager [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Terminating instance [ 1019.579191] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b45640c-92cc-4c9e-a320-8710ba9358ae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.582534] env[61985]: DEBUG nova.compute.manager [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1019.582830] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1019.586712] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472de994-0e61-48bd-8c4a-9c67e1944c3a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.597721] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a672bd6-837d-49f6-a0a6-d713e0f55d9d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.602135] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1019.602440] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11d098c7-1aa3-4be9-ae3e-47f2056669fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.618461] env[61985]: DEBUG nova.compute.provider_tree [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1019.621906] env[61985]: DEBUG oslo_vmware.api [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 1019.621906] env[61985]: value = "task-936350" [ 1019.621906] env[61985]: _type = "Task" [ 1019.621906] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.630635] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1019.632140] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffa4464-c124-4076-8577-6789f05a7485 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.638716] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936349, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.643038] env[61985]: DEBUG oslo_vmware.api [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936350, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.645289] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1019.645561] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-592843aa-9d1c-43a6-973e-a53d584a07b2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.732037] env[61985]: DEBUG nova.compute.manager [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1019.739500] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1019.739739] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1019.739949] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleting the datastore file [datastore2] afa46da0-3bd1-47ea-bdf0-28eed87fd627 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.740262] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c352f70e-b99f-48f7-8844-10fe995df670 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.747171] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1019.747171] env[61985]: value = "task-936352" [ 1019.747171] env[61985]: _type = "Task" [ 1019.747171] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.755083] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936352, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.770165] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.770482] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.770706] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.770929] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.771137] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.774236] env[61985]: INFO nova.compute.manager [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Terminating instance [ 1019.776198] env[61985]: DEBUG nova.compute.manager [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1019.776980] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1019.777426] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6cde567-5011-4c5a-a212-a46831e0bd8f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.785808] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1019.786092] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68009c4e-60e8-464c-9ddb-73089d4cebeb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.793218] env[61985]: DEBUG oslo_vmware.api [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 1019.793218] env[61985]: value = "task-936353" [ 1019.793218] env[61985]: _type = "Task" [ 1019.793218] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.804658] env[61985]: DEBUG oslo_vmware.api [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936353, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.024512] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.045139] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "49707589-4969-4e08-882b-2a2c94bc0d85" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.045507] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "49707589-4969-4e08-882b-2a2c94bc0d85" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.045803] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "49707589-4969-4e08-882b-2a2c94bc0d85-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.046010] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "49707589-4969-4e08-882b-2a2c94bc0d85-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.046198] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "49707589-4969-4e08-882b-2a2c94bc0d85-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.048582] env[61985]: INFO nova.compute.manager [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Terminating instance [ 1020.050739] env[61985]: DEBUG nova.compute.manager [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1020.050933] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1020.051765] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed6c6e6-577d-48d0-8717-5f99977feddd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.059227] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1020.059460] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ec1679c-aff4-405b-bcc1-0636261fb3d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.091099] env[61985]: DEBUG oslo_concurrency.lockutils [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.137216] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936349, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.605817} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.140402] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8db82979-e527-4c63-8ce8-fe1456f7708c/8db82979-e527-4c63-8ce8-fe1456f7708c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1020.140630] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.140905] env[61985]: DEBUG oslo_vmware.api [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936350, 'name': PowerOffVM_Task, 'duration_secs': 0.258} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.141971] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff12cf8f-d164-4a71-92b6-1097f6158978 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.143925] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1020.144118] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1020.144398] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1020.144641] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1020.144841] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleting the datastore file [datastore1] 49707589-4969-4e08-882b-2a2c94bc0d85 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.145083] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7121e5ab-e0c9-4e7a-bef5-0e3620b24899 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.146495] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f5677c7-1f76-46bb-b568-7b8cd9675d06 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.153309] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1020.153309] env[61985]: value = "task-936356" [ 1020.153309] env[61985]: _type = "Task" [ 1020.153309] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.154515] env[61985]: DEBUG oslo_vmware.api [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1020.154515] env[61985]: value = "task-936357" [ 1020.154515] env[61985]: _type = "Task" [ 1020.154515] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.158041] env[61985]: DEBUG nova.scheduler.client.report [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Updated inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 with generation 123 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1020.158332] env[61985]: DEBUG nova.compute.provider_tree [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Updating resource provider aed7e5b3-c662-4538-8447-c4f67b460215 generation from 123 to 124 during operation: update_inventory {{(pid=61985) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1020.158559] env[61985]: DEBUG nova.compute.provider_tree [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1020.171580] env[61985]: DEBUG oslo_vmware.api [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936357, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.174395] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936356, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.213672] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1020.213937] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1020.214204] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Deleting the datastore file [datastore1] 90ac7d3e-aef3-4a50-963a-606e7c58f446 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.214399] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e97d1370-7f1c-440e-a3cb-0c53f9d72e60 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.220840] env[61985]: DEBUG oslo_vmware.api [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for the task: (returnval){ [ 1020.220840] env[61985]: value = "task-936358" [ 1020.220840] env[61985]: _type = "Task" [ 1020.220840] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.228823] env[61985]: DEBUG oslo_vmware.api [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936358, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.253752] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.268095] env[61985]: DEBUG oslo_vmware.api [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936352, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258475} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.268419] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.268659] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1020.268887] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1020.288250] env[61985]: INFO nova.scheduler.client.report [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted allocations for instance afa46da0-3bd1-47ea-bdf0-28eed87fd627 [ 1020.302800] env[61985]: DEBUG oslo_vmware.api [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936353, 'name': PowerOffVM_Task, 'duration_secs': 0.222168} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.303073] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1020.303248] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1020.303482] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-91366776-841d-4db7-aa86-7b911bd3c7fd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.367039] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1020.367205] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1020.367451] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Deleting the datastore file [datastore1] cfd59b61-cca9-48d5-85e1-1f45d13f1e88 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.368149] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e1a749c-5e11-4b0e-a889-2a8c59e132db {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.375544] env[61985]: DEBUG oslo_vmware.api [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 1020.375544] env[61985]: value = "task-936360" [ 1020.375544] env[61985]: _type = "Task" [ 1020.375544] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.383972] env[61985]: DEBUG oslo_vmware.api [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936360, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.667051] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.486s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.667555] env[61985]: DEBUG nova.compute.manager [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1020.670242] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936356, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087752} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.673386] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.127s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.673618] env[61985]: DEBUG nova.objects.instance [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lazy-loading 'resources' on Instance uuid 172647b8-1f1d-49cb-a0fd-63078bc1ae85 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.674655] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.674974] env[61985]: DEBUG oslo_vmware.api [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936357, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166539} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.675796] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da41521-e4f6-4ce0-ab15-28af50a133fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.678196] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.678448] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1020.682018] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1020.682018] env[61985]: INFO nova.compute.manager [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Took 0.63 seconds to destroy the instance on the hypervisor. [ 1020.682018] env[61985]: DEBUG oslo.service.loopingcall [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.682018] env[61985]: DEBUG nova.compute.manager [-] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.682018] env[61985]: DEBUG nova.network.neutron [-] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1020.701638] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 8db82979-e527-4c63-8ce8-fe1456f7708c/8db82979-e527-4c63-8ce8-fe1456f7708c.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.702279] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd5785b1-f3d0-4bf0-ad29-432289c98406 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.726201] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1020.726201] env[61985]: value = "task-936361" [ 1020.726201] env[61985]: _type = "Task" [ 1020.726201] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.732856] env[61985]: DEBUG oslo_vmware.api [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Task: {'id': task-936358, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203133} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.733483] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.733730] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1020.733925] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1020.734145] env[61985]: INFO nova.compute.manager [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1020.734389] env[61985]: DEBUG oslo.service.loopingcall [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.734896] env[61985]: DEBUG nova.compute.manager [-] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.734994] env[61985]: DEBUG nova.network.neutron [-] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1020.739433] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936361, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.793445] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.888604] env[61985]: DEBUG oslo_vmware.api [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936360, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198955} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.890196] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.890196] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1020.890196] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1020.890196] env[61985]: INFO nova.compute.manager [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1020.890196] env[61985]: DEBUG oslo.service.loopingcall [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.890473] env[61985]: DEBUG nova.compute.manager [-] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.890576] env[61985]: DEBUG nova.network.neutron [-] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1021.071796] env[61985]: DEBUG nova.compute.manager [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Received event network-vif-unplugged-8e692ff4-8a38-44ec-ac4b-05e0bda78308 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1021.072048] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] Acquiring lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.072249] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.072435] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.072612] env[61985]: DEBUG nova.compute.manager [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] No waiting events found dispatching network-vif-unplugged-8e692ff4-8a38-44ec-ac4b-05e0bda78308 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1021.072787] env[61985]: WARNING nova.compute.manager [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Received unexpected event network-vif-unplugged-8e692ff4-8a38-44ec-ac4b-05e0bda78308 for instance with vm_state shelved_offloaded and task_state None. [ 1021.072962] env[61985]: DEBUG nova.compute.manager [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Received event network-changed-8e692ff4-8a38-44ec-ac4b-05e0bda78308 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1021.073140] env[61985]: DEBUG nova.compute.manager [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Refreshing instance network info cache due to event network-changed-8e692ff4-8a38-44ec-ac4b-05e0bda78308. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1021.073327] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] Acquiring lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.073468] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] Acquired lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.073626] env[61985]: DEBUG nova.network.neutron [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Refreshing network info cache for port 8e692ff4-8a38-44ec-ac4b-05e0bda78308 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1021.175643] env[61985]: DEBUG nova.compute.utils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1021.179886] env[61985]: DEBUG nova.compute.manager [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1021.180060] env[61985]: DEBUG nova.network.neutron [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1021.227398] env[61985]: DEBUG nova.policy [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f8df9858fb34d4cb63c2d735eb73319', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '697ed07c609f4e1f86d317675a4749a8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1021.240209] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936361, 'name': ReconfigVM_Task, 'duration_secs': 0.287852} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.240545] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 8db82979-e527-4c63-8ce8-fe1456f7708c/8db82979-e527-4c63-8ce8-fe1456f7708c.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.241300] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d0ff3842-956d-4ee9-8158-d202d094f636 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.248063] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1021.248063] env[61985]: value = "task-936362" [ 1021.248063] env[61985]: _type = "Task" [ 1021.248063] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.257196] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936362, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.457973] env[61985]: DEBUG nova.network.neutron [-] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.482550] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94517b38-7d1b-4832-8be4-784002ffd795 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.490478] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3fc97fb-a59d-4c75-8c2a-d29c1a2b67bf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.495366] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "87006fbf-ea90-4d9a-88af-001de424ac14" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.495543] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "87006fbf-ea90-4d9a-88af-001de424ac14" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.495730] env[61985]: INFO nova.compute.manager [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Shelving [ 1021.526493] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d7de22-b151-45f7-b4d4-68a68ce55008 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.530156] env[61985]: DEBUG nova.compute.manager [req-24383c5a-bef5-4861-841e-bfdafa355e69 req-39283d66-c3a1-4bf5-bfe1-1dd0b96bcbd5 service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Received event network-vif-deleted-63023688-9d9e-4fab-91af-9e9171bccc16 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1021.530328] env[61985]: INFO nova.compute.manager [req-24383c5a-bef5-4861-841e-bfdafa355e69 req-39283d66-c3a1-4bf5-bfe1-1dd0b96bcbd5 service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Neutron deleted interface 63023688-9d9e-4fab-91af-9e9171bccc16; detaching it from the instance and deleting it from the info cache [ 1021.530506] env[61985]: DEBUG nova.network.neutron [req-24383c5a-bef5-4861-841e-bfdafa355e69 req-39283d66-c3a1-4bf5-bfe1-1dd0b96bcbd5 service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.539644] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93132fa0-ca60-4486-b0e6-23c813594622 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.555501] env[61985]: DEBUG nova.compute.provider_tree [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.582879] env[61985]: DEBUG nova.network.neutron [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Successfully created port: a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1021.684496] env[61985]: DEBUG nova.compute.manager [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1021.757831] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936362, 'name': Rename_Task, 'duration_secs': 0.132543} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.758158] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1021.758410] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b27ea51-5175-4083-b6e7-12623732be70 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.763918] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1021.763918] env[61985]: value = "task-936363" [ 1021.763918] env[61985]: _type = "Task" [ 1021.763918] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.766713] env[61985]: DEBUG nova.network.neutron [-] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.772418] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936363, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.869114] env[61985]: DEBUG nova.network.neutron [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Updated VIF entry in instance network info cache for port 8e692ff4-8a38-44ec-ac4b-05e0bda78308. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1021.869519] env[61985]: DEBUG nova.network.neutron [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Updating instance_info_cache with network_info: [{"id": "8e692ff4-8a38-44ec-ac4b-05e0bda78308", "address": "fa:16:3e:80:98:79", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": null, "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap8e692ff4-8a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.911946] env[61985]: DEBUG nova.network.neutron [-] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.960153] env[61985]: INFO nova.compute.manager [-] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Took 1.28 seconds to deallocate network for instance. [ 1022.002543] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.002832] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d9ae195-6094-429b-8d8e-a23237d9ef1c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.010652] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1022.010652] env[61985]: value = "task-936364" [ 1022.010652] env[61985]: _type = "Task" [ 1022.010652] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.019056] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.035337] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee71effe-50a6-466c-9aaa-a1a2a8455b45 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.044550] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13a4729-ec38-4026-9db8-9ddfc0054afa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.060419] env[61985]: DEBUG nova.scheduler.client.report [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.074850] env[61985]: DEBUG nova.compute.manager [req-24383c5a-bef5-4861-841e-bfdafa355e69 req-39283d66-c3a1-4bf5-bfe1-1dd0b96bcbd5 service nova] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Detach interface failed, port_id=63023688-9d9e-4fab-91af-9e9171bccc16, reason: Instance 90ac7d3e-aef3-4a50-963a-606e7c58f446 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1022.150385] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.269481] env[61985]: INFO nova.compute.manager [-] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Took 1.53 seconds to deallocate network for instance. [ 1022.278053] env[61985]: DEBUG oslo_vmware.api [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936363, 'name': PowerOnVM_Task, 'duration_secs': 0.439549} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.278319] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1022.278495] env[61985]: INFO nova.compute.manager [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Took 9.53 seconds to spawn the instance on the hypervisor. [ 1022.278686] env[61985]: DEBUG nova.compute.manager [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.279470] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dece689-eb54-4060-bdbc-ec518b72e9a5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.372253] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a87c40f-10a5-4e51-93b6-3d106d35c9a9 req-fe649d3a-9ace-496b-a047-8a983db2370f service nova] Releasing lock "refresh_cache-afa46da0-3bd1-47ea-bdf0-28eed87fd627" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.414108] env[61985]: INFO nova.compute.manager [-] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Took 1.52 seconds to deallocate network for instance. [ 1022.466641] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.520277] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936364, 'name': PowerOffVM_Task, 'duration_secs': 0.21168} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.520559] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1022.521352] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0449fb45-0f31-4282-a383-e8e4b0860059 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.539542] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40375ba-b371-4424-aa2a-9568028f8a1c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.565525] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.892s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.567810] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.273s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.569341] env[61985]: INFO nova.compute.claims [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1022.587946] env[61985]: INFO nova.scheduler.client.report [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleted allocations for instance 172647b8-1f1d-49cb-a0fd-63078bc1ae85 [ 1022.695181] env[61985]: DEBUG nova.compute.manager [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1022.720948] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1022.721301] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1022.721389] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.721578] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1022.721752] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.721906] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1022.722250] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1022.722375] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1022.723501] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1022.723501] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1022.723501] env[61985]: DEBUG nova.virt.hardware [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1022.724090] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b4beaa-9e81-4d60-96f3-4bde1c25953d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.733674] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70855f7-b1c0-4e1d-a582-c4880ef00a2d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.780171] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.795590] env[61985]: INFO nova.compute.manager [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Took 29.72 seconds to build instance. [ 1022.921414] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.053178] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1023.053178] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6a2518ed-e6f6-4467-b27b-135bb6e9958a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.064377] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1023.064377] env[61985]: value = "task-936365" [ 1023.064377] env[61985]: _type = "Task" [ 1023.064377] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.071759] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936365, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.095345] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c7b399fc-a918-49a1-8fec-3188b81b1212 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "172647b8-1f1d-49cb-a0fd-63078bc1ae85" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.291s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.111241] env[61985]: DEBUG nova.compute.manager [req-2dec95e1-f27b-494f-a05a-0ac8a0358c42 req-f72c4c97-aac4-4bed-95ba-be78ad847521 service nova] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Received event network-vif-deleted-4b48329e-7fd7-4641-845a-68423e22587b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1023.297869] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e1f4da2-ce3c-4fac-9655-babcba858a8c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "8db82979-e527-4c63-8ce8-fe1456f7708c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.232s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.499363] env[61985]: DEBUG nova.network.neutron [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Successfully updated port: a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1023.580414] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936365, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.587161] env[61985]: DEBUG nova.compute.manager [req-badc68e8-ad9a-4d40-bbf0-2c3f1357dbd3 req-cb3d1b13-f6d8-49e9-a18e-a5bd96e901f1 service nova] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Received event network-vif-deleted-c529f7f6-f908-4a69-944f-b979f622f99f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1023.849094] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604d402d-c2d8-464d-b4aa-e0f8e9cb3514 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.859200] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e527e4-c951-43e0-8f38-8a0a135c8cae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.894009] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2629f707-f3a6-4d64-9484-73b437f2dc7d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.901716] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb559e1f-f5e9-4932-8778-0629fa2c4b5d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.917860] env[61985]: DEBUG nova.compute.provider_tree [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.005631] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.005936] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.006030] env[61985]: DEBUG nova.network.neutron [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1024.073442] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936365, 'name': CreateSnapshot_Task, 'duration_secs': 0.576732} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.073727] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1024.074471] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec5aa7a-36a1-45b4-9bd3-85d60fc12555 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.424029] env[61985]: DEBUG nova.scheduler.client.report [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.436114] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.436624] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.538081] env[61985]: DEBUG nova.network.neutron [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1024.591955] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1024.592485] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e8caeb65-d432-4caa-b26a-62ecaba852b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.601352] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1024.601352] env[61985]: value = "task-936366" [ 1024.601352] env[61985]: _type = "Task" [ 1024.601352] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.612738] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936366, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.689778] env[61985]: DEBUG nova.network.neutron [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance_info_cache with network_info: [{"id": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "address": "fa:16:3e:e1:d0:1f", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64c525f-b3", "ovs_interfaceid": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.927115] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.927673] env[61985]: DEBUG nova.compute.manager [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1024.931687] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.747s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.932195] env[61985]: DEBUG nova.objects.instance [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lazy-loading 'resources' on Instance uuid 986c576e-be02-48ac-b24c-72edccab25be {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.942697] env[61985]: INFO nova.compute.manager [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Detaching volume 774d6c9b-3a7e-48dc-836f-82cec83d6e53 [ 1024.986049] env[61985]: INFO nova.virt.block_device [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Attempting to driver detach volume 774d6c9b-3a7e-48dc-836f-82cec83d6e53 from mountpoint /dev/sdb [ 1024.986338] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1024.986554] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211414', 'volume_id': '774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'name': 'volume-774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce757dda-f58a-47b3-b319-e6b01f05b20b', 'attached_at': '', 'detached_at': '', 'volume_id': '774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'serial': '774d6c9b-3a7e-48dc-836f-82cec83d6e53'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1024.987505] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a63f60-2326-4156-ad00-2e3ce4a92ab7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.021070] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26f2324-c5dd-49a0-9c85-212ab52f9f53 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.027928] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d416cd08-f706-4fb3-a4d8-413812e1b2d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.051314] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776e916d-2ef8-4d0f-8d9b-0381a4ee11dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.070586] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] The volume has not been displaced from its original location: [datastore1] volume-774d6c9b-3a7e-48dc-836f-82cec83d6e53/volume-774d6c9b-3a7e-48dc-836f-82cec83d6e53.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1025.076472] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Reconfiguring VM instance instance-0000003d to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1025.077250] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1616c41-e9b1-4aef-a43a-2ec9ba35200d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.097199] env[61985]: DEBUG oslo_vmware.api [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1025.097199] env[61985]: value = "task-936367" [ 1025.097199] env[61985]: _type = "Task" [ 1025.097199] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.108620] env[61985]: DEBUG oslo_vmware.api [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936367, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.114298] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936366, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.148252] env[61985]: DEBUG nova.compute.manager [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Received event network-vif-plugged-a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1025.148517] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Acquiring lock "885a36d9-24c6-407a-9d6b-01cd97e90674-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.148747] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.148947] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.149218] env[61985]: DEBUG nova.compute.manager [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] No waiting events found dispatching network-vif-plugged-a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1025.149458] env[61985]: WARNING nova.compute.manager [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Received unexpected event network-vif-plugged-a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 for instance with vm_state building and task_state spawning. [ 1025.149655] env[61985]: DEBUG nova.compute.manager [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Received event network-changed-a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1025.149905] env[61985]: DEBUG nova.compute.manager [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Refreshing instance network info cache due to event network-changed-a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1025.150243] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Acquiring lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.193083] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.193492] env[61985]: DEBUG nova.compute.manager [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Instance network_info: |[{"id": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "address": "fa:16:3e:e1:d0:1f", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64c525f-b3", "ovs_interfaceid": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1025.193854] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Acquired lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.194080] env[61985]: DEBUG nova.network.neutron [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Refreshing network info cache for port a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1025.195735] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:d0:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.205202] env[61985]: DEBUG oslo.service.loopingcall [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.209856] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1025.210689] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-acd17035-3a98-4d1d-8201-e0b25dea922b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.245933] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.245933] env[61985]: value = "task-936368" [ 1025.245933] env[61985]: _type = "Task" [ 1025.245933] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.259260] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936368, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.435255] env[61985]: DEBUG nova.compute.utils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1025.436910] env[61985]: DEBUG nova.compute.manager [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1025.437306] env[61985]: DEBUG nova.network.neutron [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1025.490438] env[61985]: DEBUG nova.network.neutron [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updated VIF entry in instance network info cache for port a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1025.490869] env[61985]: DEBUG nova.network.neutron [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance_info_cache with network_info: [{"id": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "address": "fa:16:3e:e1:d0:1f", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64c525f-b3", "ovs_interfaceid": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.542953] env[61985]: DEBUG nova.policy [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da7c7d5d10b34b4ebdede5ff3cfd8b01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e66b7a6e7354b06a77295eadf4e5e0f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1025.617304] env[61985]: DEBUG oslo_vmware.api [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936367, 'name': ReconfigVM_Task, 'duration_secs': 0.220795} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.622538] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Reconfigured VM instance instance-0000003d to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1025.627324] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936366, 'name': CloneVM_Task, 'duration_secs': 1.013677} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.627951] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0affb36-b2f0-4e5a-bf46-d78bc750c2f4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.637990] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Created linked-clone VM from snapshot [ 1025.641636] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20dcdd34-9d10-452d-a04d-0945414525c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.645452] env[61985]: DEBUG nova.compute.manager [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Received event network-changed-88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1025.645666] env[61985]: DEBUG nova.compute.manager [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Refreshing instance network info cache due to event network-changed-88a63ff6-55f6-4b8b-ba9e-188245f7ea87. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1025.645891] env[61985]: DEBUG oslo_concurrency.lockutils [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] Acquiring lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.646051] env[61985]: DEBUG oslo_concurrency.lockutils [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] Acquired lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.646222] env[61985]: DEBUG nova.network.neutron [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Refreshing network info cache for port 88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1025.653780] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Uploading image 173db284-6b8b-4355-bdde-6a9ba3be888e {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1025.657887] env[61985]: DEBUG oslo_vmware.api [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1025.657887] env[61985]: value = "task-936369" [ 1025.657887] env[61985]: _type = "Task" [ 1025.657887] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.669037] env[61985]: DEBUG oslo_vmware.api [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936369, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.682125] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1025.682125] env[61985]: value = "vm-211449" [ 1025.682125] env[61985]: _type = "VirtualMachine" [ 1025.682125] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1025.682436] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3db27f60-71d0-4668-92e4-c3051370354d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.691452] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lease: (returnval){ [ 1025.691452] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523fc1f9-b49a-d16f-f351-1cd93afc9c56" [ 1025.691452] env[61985]: _type = "HttpNfcLease" [ 1025.691452] env[61985]: } obtained for exporting VM: (result){ [ 1025.691452] env[61985]: value = "vm-211449" [ 1025.691452] env[61985]: _type = "VirtualMachine" [ 1025.691452] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1025.691777] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the lease: (returnval){ [ 1025.691777] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523fc1f9-b49a-d16f-f351-1cd93afc9c56" [ 1025.691777] env[61985]: _type = "HttpNfcLease" [ 1025.691777] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1025.702295] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1025.702295] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523fc1f9-b49a-d16f-f351-1cd93afc9c56" [ 1025.702295] env[61985]: _type = "HttpNfcLease" [ 1025.702295] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1025.736410] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d43a92-994c-4ed4-aa56-937534919995 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.744708] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c005df4-223a-4224-a07e-ff97f9404653 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.756096] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936368, 'name': CreateVM_Task, 'duration_secs': 0.348481} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.780348] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1025.781350] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.781556] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.781927] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1025.782658] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989f4b22-d230-4317-9457-eac2ab4515dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.785256] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16eda1b3-f7ab-4adb-9946-b40e08620b1d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.792035] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1025.792035] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526a271a-b6c2-e57c-0e4e-8cb817f960ca" [ 1025.792035] env[61985]: _type = "Task" [ 1025.792035] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.798305] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87bc63f8-1009-4374-8022-67f3768aa1eb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.806782] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526a271a-b6c2-e57c-0e4e-8cb817f960ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.814641] env[61985]: DEBUG nova.compute.provider_tree [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.878979] env[61985]: DEBUG nova.network.neutron [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Successfully created port: 6d05ecf8-3c82-464f-be11-a7e2766f9351 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.940106] env[61985]: DEBUG nova.compute.manager [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1025.995631] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Releasing lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.995918] env[61985]: DEBUG nova.compute.manager [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received event network-changed-7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1025.996113] env[61985]: DEBUG nova.compute.manager [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing instance network info cache due to event network-changed-7075efce-177a-4127-9c2f-7e3a9d1e92ad. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1025.996328] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Acquiring lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.996479] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Acquired lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.996643] env[61985]: DEBUG nova.network.neutron [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing network info cache for port 7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1026.086304] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "interface-5f2aa808-8d1e-471e-9ef7-0f91590ec546-0822504d-b73d-40b5-a7a2-4770f9d2f190" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.086581] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5f2aa808-8d1e-471e-9ef7-0f91590ec546-0822504d-b73d-40b5-a7a2-4770f9d2f190" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.086970] env[61985]: DEBUG nova.objects.instance [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'flavor' on Instance uuid 5f2aa808-8d1e-471e-9ef7-0f91590ec546 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.179019] env[61985]: DEBUG oslo_vmware.api [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936369, 'name': ReconfigVM_Task, 'duration_secs': 0.144376} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.179019] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211414', 'volume_id': '774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'name': 'volume-774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce757dda-f58a-47b3-b319-e6b01f05b20b', 'attached_at': '', 'detached_at': '', 'volume_id': '774d6c9b-3a7e-48dc-836f-82cec83d6e53', 'serial': '774d6c9b-3a7e-48dc-836f-82cec83d6e53'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1026.204348] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1026.204348] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523fc1f9-b49a-d16f-f351-1cd93afc9c56" [ 1026.204348] env[61985]: _type = "HttpNfcLease" [ 1026.204348] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1026.204704] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1026.204704] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523fc1f9-b49a-d16f-f351-1cd93afc9c56" [ 1026.204704] env[61985]: _type = "HttpNfcLease" [ 1026.204704] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1026.205475] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5713c1d-2f70-4fe0-bf6f-9621cdb91f08 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.213490] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527e3c7d-abd8-2dd4-bb92-6f207d0d7d26/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1026.213683] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527e3c7d-abd8-2dd4-bb92-6f207d0d7d26/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1026.304430] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526a271a-b6c2-e57c-0e4e-8cb817f960ca, 'name': SearchDatastore_Task, 'duration_secs': 0.031139} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.304717] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.305158] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.305508] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.305691] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.305890] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.306208] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75950cf8-1724-4975-a1d3-652445989a63 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.311048] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ae3c0cda-1c1e-4456-8778-0a8b02305ff3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.316638] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1026.316846] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1026.318217] env[61985]: DEBUG nova.scheduler.client.report [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.321370] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2651343f-5737-4007-bdba-59d7657f7db5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.327346] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1026.327346] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520980ca-c907-2478-3a50-1d86b8b89c58" [ 1026.327346] env[61985]: _type = "Task" [ 1026.327346] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.334832] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520980ca-c907-2478-3a50-1d86b8b89c58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.419111] env[61985]: DEBUG nova.network.neutron [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updated VIF entry in instance network info cache for port 88a63ff6-55f6-4b8b-ba9e-188245f7ea87. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1026.419316] env[61985]: DEBUG nova.network.neutron [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updating instance_info_cache with network_info: [{"id": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "address": "fa:16:3e:16:ec:4d", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88a63ff6-55", "ovs_interfaceid": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.715023] env[61985]: DEBUG nova.network.neutron [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updated VIF entry in instance network info cache for port 7075efce-177a-4127-9c2f-7e3a9d1e92ad. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1026.715814] env[61985]: DEBUG nova.network.neutron [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [{"id": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "address": "fa:16:3e:4b:3d:56", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7075efce-17", "ovs_interfaceid": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.735581] env[61985]: DEBUG nova.objects.instance [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'pci_requests' on Instance uuid 5f2aa808-8d1e-471e-9ef7-0f91590ec546 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.740880] env[61985]: DEBUG nova.objects.instance [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lazy-loading 'flavor' on Instance uuid ce757dda-f58a-47b3-b319-e6b01f05b20b {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.825807] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.829890] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.498s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.829890] env[61985]: DEBUG nova.objects.instance [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lazy-loading 'resources' on Instance uuid bed26b5f-7ca5-405a-884e-02b2495dd977 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.841549] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520980ca-c907-2478-3a50-1d86b8b89c58, 'name': SearchDatastore_Task, 'duration_secs': 0.028502} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.842709] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73d15f45-4896-4ceb-a7a4-0938a22d7acf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.848974] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1026.848974] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523e1d37-408b-32a0-c07e-b1704a0b7966" [ 1026.848974] env[61985]: _type = "Task" [ 1026.848974] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.853419] env[61985]: INFO nova.scheduler.client.report [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleted allocations for instance 986c576e-be02-48ac-b24c-72edccab25be [ 1026.861537] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523e1d37-408b-32a0-c07e-b1704a0b7966, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.923421] env[61985]: DEBUG oslo_concurrency.lockutils [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] Releasing lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.923748] env[61985]: DEBUG nova.compute.manager [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received event network-changed-7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1026.923936] env[61985]: DEBUG nova.compute.manager [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing instance network info cache due to event network-changed-7075efce-177a-4127-9c2f-7e3a9d1e92ad. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1026.924169] env[61985]: DEBUG oslo_concurrency.lockutils [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] Acquiring lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.950662] env[61985]: DEBUG nova.compute.manager [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1026.978322] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.978567] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.978760] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.979111] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.979349] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.979555] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.979815] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.980054] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.980307] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.980576] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.980853] env[61985]: DEBUG nova.virt.hardware [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.981815] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b59a7fc-cf7b-425b-a154-d9acca278379 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.990244] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231f93fe-94d3-4d9c-a5db-d7aeb006e295 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.218510] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Releasing lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.218905] env[61985]: DEBUG nova.compute.manager [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Received event network-changed-88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1027.219109] env[61985]: DEBUG nova.compute.manager [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Refreshing instance network info cache due to event network-changed-88a63ff6-55f6-4b8b-ba9e-188245f7ea87. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1027.219340] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Acquiring lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.219492] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Acquired lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.219660] env[61985]: DEBUG nova.network.neutron [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Refreshing network info cache for port 88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1027.220914] env[61985]: DEBUG oslo_concurrency.lockutils [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] Acquired lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.221118] env[61985]: DEBUG nova.network.neutron [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing network info cache for port 7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1027.244053] env[61985]: DEBUG nova.objects.base [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Object Instance<5f2aa808-8d1e-471e-9ef7-0f91590ec546> lazy-loaded attributes: flavor,pci_requests {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1027.244300] env[61985]: DEBUG nova.network.neutron [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1027.372243] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1285f768-990b-4ab5-8204-bcf9439e5579 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "986c576e-be02-48ac-b24c-72edccab25be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.319s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.375039] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523e1d37-408b-32a0-c07e-b1704a0b7966, 'name': SearchDatastore_Task, 'duration_secs': 0.017161} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.378040] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.378373] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 885a36d9-24c6-407a-9d6b-01cd97e90674/885a36d9-24c6-407a-9d6b-01cd97e90674.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1027.379012] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b1d4fb1-a637-47e0-9779-07699cecf73b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.382819] env[61985]: DEBUG nova.policy [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a78dbc667d24c819e6447740fc2d8e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2dbceebf07c43ac90ed80f8a09294b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1027.389538] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1027.389538] env[61985]: value = "task-936371" [ 1027.389538] env[61985]: _type = "Task" [ 1027.389538] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.408708] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936371, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.600753] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f146d9-40e3-4ac5-9753-41be497ac88c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.611640] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24c6d48-25b7-46c0-b041-35d385d10128 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.648873] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4856137a-0ab1-4a68-8ab8-89092254bb75 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.658243] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e4221d-9541-473e-a14c-7d782aee4cd1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.674253] env[61985]: DEBUG nova.compute.provider_tree [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.752817] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d0331b1-58a3-4d85-98ed-17f45a629d90 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.316s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.854673] env[61985]: DEBUG nova.network.neutron [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Successfully updated port: 6d05ecf8-3c82-464f-be11-a7e2766f9351 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1027.902952] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936371, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.072406] env[61985]: DEBUG nova.compute.manager [req-1ec27f80-243f-45a6-86ed-801d32eeb0fe req-d447c1e7-bd3c-4059-95f3-97d686f2f83f service nova] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Received event network-vif-plugged-6d05ecf8-3c82-464f-be11-a7e2766f9351 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1028.072501] env[61985]: DEBUG oslo_concurrency.lockutils [req-1ec27f80-243f-45a6-86ed-801d32eeb0fe req-d447c1e7-bd3c-4059-95f3-97d686f2f83f service nova] Acquiring lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.072739] env[61985]: DEBUG oslo_concurrency.lockutils [req-1ec27f80-243f-45a6-86ed-801d32eeb0fe req-d447c1e7-bd3c-4059-95f3-97d686f2f83f service nova] Lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.072957] env[61985]: DEBUG oslo_concurrency.lockutils [req-1ec27f80-243f-45a6-86ed-801d32eeb0fe req-d447c1e7-bd3c-4059-95f3-97d686f2f83f service nova] Lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.073194] env[61985]: DEBUG nova.compute.manager [req-1ec27f80-243f-45a6-86ed-801d32eeb0fe req-d447c1e7-bd3c-4059-95f3-97d686f2f83f service nova] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] No waiting events found dispatching network-vif-plugged-6d05ecf8-3c82-464f-be11-a7e2766f9351 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1028.073418] env[61985]: WARNING nova.compute.manager [req-1ec27f80-243f-45a6-86ed-801d32eeb0fe req-d447c1e7-bd3c-4059-95f3-97d686f2f83f service nova] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Received unexpected event network-vif-plugged-6d05ecf8-3c82-464f-be11-a7e2766f9351 for instance with vm_state building and task_state spawning. [ 1028.085036] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.085281] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.085598] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "ce757dda-f58a-47b3-b319-e6b01f05b20b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.085698] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.085865] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.087971] env[61985]: INFO nova.compute.manager [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Terminating instance [ 1028.090373] env[61985]: DEBUG nova.compute.manager [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1028.090624] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1028.096879] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbf3778-36da-4813-869b-7c2d3c10f966 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.106851] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1028.108304] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abdb541b-0dee-4b6f-94e3-32a351774be9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.119923] env[61985]: DEBUG oslo_vmware.api [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1028.119923] env[61985]: value = "task-936372" [ 1028.119923] env[61985]: _type = "Task" [ 1028.119923] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.134717] env[61985]: DEBUG oslo_vmware.api [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936372, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.173597] env[61985]: DEBUG nova.network.neutron [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updated VIF entry in instance network info cache for port 88a63ff6-55f6-4b8b-ba9e-188245f7ea87. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1028.174207] env[61985]: DEBUG nova.network.neutron [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updating instance_info_cache with network_info: [{"id": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "address": "fa:16:3e:16:ec:4d", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88a63ff6-55", "ovs_interfaceid": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.178308] env[61985]: DEBUG nova.scheduler.client.report [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.183387] env[61985]: DEBUG nova.network.neutron [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updated VIF entry in instance network info cache for port 7075efce-177a-4127-9c2f-7e3a9d1e92ad. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1028.183839] env[61985]: DEBUG nova.network.neutron [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [{"id": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "address": "fa:16:3e:4b:3d:56", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7075efce-17", "ovs_interfaceid": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.357559] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "refresh_cache-5cd5e044-b0da-4564-8c2c-e894eb29a74c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.357810] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "refresh_cache-5cd5e044-b0da-4564-8c2c-e894eb29a74c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.357992] env[61985]: DEBUG nova.network.neutron [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1028.403058] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936371, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.65608} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.403223] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 885a36d9-24c6-407a-9d6b-01cd97e90674/885a36d9-24c6-407a-9d6b-01cd97e90674.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1028.403377] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1028.403649] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-781551db-f7dc-450e-b672-be4ac730ce6a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.410926] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1028.410926] env[61985]: value = "task-936373" [ 1028.410926] env[61985]: _type = "Task" [ 1028.410926] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.418785] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936373, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.631319] env[61985]: DEBUG oslo_vmware.api [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936372, 'name': PowerOffVM_Task, 'duration_secs': 0.197306} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.631698] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1028.631888] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1028.632208] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc8cd3c2-4d70-429c-af2a-3e35065c9826 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.678009] env[61985]: DEBUG oslo_concurrency.lockutils [req-75f2bcf4-c84e-44f9-9f69-9e876ccb26e5 req-71699eb9-302b-4e0c-808b-51c23c43f94c service nova] Releasing lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.688054] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.858s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.689975] env[61985]: DEBUG oslo_concurrency.lockutils [req-96b41671-d6a7-4a65-95bd-6ad60b9f27fb req-15e9edba-9ed4-4941-a5a5-a304c53962b4 service nova] Releasing lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.690724] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.298s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.692650] env[61985]: INFO nova.compute.claims [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.711326] env[61985]: INFO nova.scheduler.client.report [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Deleted allocations for instance bed26b5f-7ca5-405a-884e-02b2495dd977 [ 1028.891251] env[61985]: DEBUG nova.network.neutron [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1028.920155] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936373, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065192} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.920423] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.921242] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541464e6-0351-4f8f-b420-2a09e7d844ad {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.943432] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 885a36d9-24c6-407a-9d6b-01cd97e90674/885a36d9-24c6-407a-9d6b-01cd97e90674.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.945928] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-facfb5a3-c754-457c-aa26-35d8590834d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.965775] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1028.965775] env[61985]: value = "task-936375" [ 1028.965775] env[61985]: _type = "Task" [ 1028.965775] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.973916] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936375, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.068649] env[61985]: DEBUG nova.network.neutron [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Updating instance_info_cache with network_info: [{"id": "6d05ecf8-3c82-464f-be11-a7e2766f9351", "address": "fa:16:3e:eb:4e:cd", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d05ecf8-3c", "ovs_interfaceid": "6d05ecf8-3c82-464f-be11-a7e2766f9351", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.220988] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2614f911-ff4b-4e45-b74d-dade00d468be tempest-ServerTagsTestJSON-1644827708 tempest-ServerTagsTestJSON-1644827708-project-member] Lock "bed26b5f-7ca5-405a-884e-02b2495dd977" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.619s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.301192] env[61985]: DEBUG nova.network.neutron [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Successfully updated port: 0822504d-b73d-40b5-a7a2-4770f9d2f190 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1029.477704] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936375, 'name': ReconfigVM_Task, 'duration_secs': 0.40496} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.478043] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 885a36d9-24c6-407a-9d6b-01cd97e90674/885a36d9-24c6-407a-9d6b-01cd97e90674.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.478657] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a62b1ba-740d-4dae-91c0-6f0ff00349b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.484928] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1029.484928] env[61985]: value = "task-936376" [ 1029.484928] env[61985]: _type = "Task" [ 1029.484928] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.493099] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936376, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.572125] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "refresh_cache-5cd5e044-b0da-4564-8c2c-e894eb29a74c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.572472] env[61985]: DEBUG nova.compute.manager [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Instance network_info: |[{"id": "6d05ecf8-3c82-464f-be11-a7e2766f9351", "address": "fa:16:3e:eb:4e:cd", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d05ecf8-3c", "ovs_interfaceid": "6d05ecf8-3c82-464f-be11-a7e2766f9351", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1029.572941] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:4e:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2c019b6-3ef3-4c8f-95bd-edede2c554a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d05ecf8-3c82-464f-be11-a7e2766f9351', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.581269] env[61985]: DEBUG oslo.service.loopingcall [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.581957] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1029.582260] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ee6f3c1-b96d-4da2-9218-8a0bcb92967e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.602851] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.602851] env[61985]: value = "task-936377" [ 1029.602851] env[61985]: _type = "Task" [ 1029.602851] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.610708] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936377, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.628857] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1029.629150] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1029.629314] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleting the datastore file [datastore1] ce757dda-f58a-47b3-b319-e6b01f05b20b {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.629597] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc75baed-78bb-49d7-b774-5015c55a4fa8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.637449] env[61985]: DEBUG oslo_vmware.api [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1029.637449] env[61985]: value = "task-936378" [ 1029.637449] env[61985]: _type = "Task" [ 1029.637449] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.646120] env[61985]: DEBUG oslo_vmware.api [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936378, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.807410] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.807644] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.807808] env[61985]: DEBUG nova.network.neutron [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1030.180118] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936377, 'name': CreateVM_Task, 'duration_secs': 0.387264} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.180418] env[61985]: DEBUG oslo_vmware.api [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936378, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219335} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.183525] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1030.183862] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1030.184070] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1030.184260] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1030.184441] env[61985]: INFO nova.compute.manager [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Took 2.09 seconds to destroy the instance on the hypervisor. [ 1030.184730] env[61985]: DEBUG oslo.service.loopingcall [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.185335] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936376, 'name': Rename_Task, 'duration_secs': 0.192923} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.188824] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.189097] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.189401] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1030.189688] env[61985]: DEBUG nova.compute.manager [-] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1030.190570] env[61985]: DEBUG nova.network.neutron [-] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1030.191786] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1030.193113] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab5e0396-b30a-416b-9c4d-5b91ef6988c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.194669] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9744b586-e7dc-495f-88e3-a50fc6f25b84 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.200562] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1030.200562] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52425dfe-ef87-811f-bc3b-79ad18fe2d58" [ 1030.200562] env[61985]: _type = "Task" [ 1030.200562] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.206033] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1030.206033] env[61985]: value = "task-936379" [ 1030.206033] env[61985]: _type = "Task" [ 1030.206033] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.218822] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52425dfe-ef87-811f-bc3b-79ad18fe2d58, 'name': SearchDatastore_Task, 'duration_secs': 0.011807} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.219789] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.220158] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1030.220588] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.220758] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.221099] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1030.230150] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc20d80a-0d0a-4e6f-aef2-223e9fb504cd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.232281] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936379, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.241074] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1030.241257] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1030.242140] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-982bbe1c-0c84-45d0-a2ae-df3a43d5ce7f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.253385] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1030.253385] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5207981f-000e-0289-d29e-20578232c1c3" [ 1030.253385] env[61985]: _type = "Task" [ 1030.253385] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.267389] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5207981f-000e-0289-d29e-20578232c1c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.356676] env[61985]: WARNING nova.network.neutron [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] 5014dbd4-285f-43e0-ad95-2107c34e808a already exists in list: networks containing: ['5014dbd4-285f-43e0-ad95-2107c34e808a']. ignoring it [ 1030.379924] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87556e53-939d-4bdc-bbd1-03828b0852fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.388592] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fb9890f-278b-42d4-ba1d-f45b05b07353 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.424882] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5948fa7c-75b9-463a-a9c2-ca1d03791aa4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.432907] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7199c725-52e3-4db9-85be-2b5c5bc48a00 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.446842] env[61985]: DEBUG nova.compute.provider_tree [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.688686] env[61985]: DEBUG nova.compute.manager [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Received event network-changed-6d05ecf8-3c82-464f-be11-a7e2766f9351 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1030.688686] env[61985]: DEBUG nova.compute.manager [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Refreshing instance network info cache due to event network-changed-6d05ecf8-3c82-464f-be11-a7e2766f9351. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1030.688686] env[61985]: DEBUG oslo_concurrency.lockutils [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] Acquiring lock "refresh_cache-5cd5e044-b0da-4564-8c2c-e894eb29a74c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.689047] env[61985]: DEBUG oslo_concurrency.lockutils [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] Acquired lock "refresh_cache-5cd5e044-b0da-4564-8c2c-e894eb29a74c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.689047] env[61985]: DEBUG nova.network.neutron [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Refreshing network info cache for port 6d05ecf8-3c82-464f-be11-a7e2766f9351 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1030.717051] env[61985]: DEBUG nova.network.neutron [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [{"id": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "address": "fa:16:3e:4b:3d:56", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7075efce-17", "ovs_interfaceid": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0822504d-b73d-40b5-a7a2-4770f9d2f190", "address": "fa:16:3e:ab:c2:ce", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0822504d-b7", "ovs_interfaceid": "0822504d-b73d-40b5-a7a2-4770f9d2f190", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.726875] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936379, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.769028] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5207981f-000e-0289-d29e-20578232c1c3, 'name': SearchDatastore_Task, 'duration_secs': 0.014402} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.769028] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82e3e7b0-4393-4f83-b1bd-80250f3c1392 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.772547] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1030.772547] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5201dfd5-620d-aa73-f776-1bf0a25ada02" [ 1030.772547] env[61985]: _type = "Task" [ 1030.772547] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.782414] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5201dfd5-620d-aa73-f776-1bf0a25ada02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.952049] env[61985]: DEBUG nova.scheduler.client.report [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.221167] env[61985]: DEBUG oslo_vmware.api [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936379, 'name': PowerOnVM_Task, 'duration_secs': 0.595328} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.224300] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1031.224580] env[61985]: INFO nova.compute.manager [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Took 8.53 seconds to spawn the instance on the hypervisor. [ 1031.224809] env[61985]: DEBUG nova.compute.manager [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1031.225681] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb5ee68-3721-4523-bddd-fd5ba6a4b996 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.228867] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.229513] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.229754] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.230629] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe5e55c-5107-4089-b6e8-7e65033efee4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.249687] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.250043] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.250360] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.250613] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.250815] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.251020] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.251280] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.251491] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.251710] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.251925] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.252158] env[61985]: DEBUG nova.virt.hardware [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.258720] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Reconfiguring VM to attach interface {{(pid=61985) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1031.260596] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d863d144-f1df-4aaa-94da-93e931fae4f3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.295899] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5201dfd5-620d-aa73-f776-1bf0a25ada02, 'name': SearchDatastore_Task, 'duration_secs': 0.009421} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.296287] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.296731] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 5cd5e044-b0da-4564-8c2c-e894eb29a74c/5cd5e044-b0da-4564-8c2c-e894eb29a74c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1031.297035] env[61985]: DEBUG oslo_vmware.api [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1031.297035] env[61985]: value = "task-936380" [ 1031.297035] env[61985]: _type = "Task" [ 1031.297035] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.297400] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36e260f7-5049-445b-b08e-171939224434 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.308269] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1031.308269] env[61985]: value = "task-936381" [ 1031.308269] env[61985]: _type = "Task" [ 1031.308269] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.312189] env[61985]: DEBUG oslo_vmware.api [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936380, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.319529] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.454939] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.764s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.455829] env[61985]: DEBUG nova.compute.manager [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1031.458267] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 13.006s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.459062] env[61985]: DEBUG nova.objects.instance [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61985) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1031.469796] env[61985]: DEBUG nova.network.neutron [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Updated VIF entry in instance network info cache for port 6d05ecf8-3c82-464f-be11-a7e2766f9351. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1031.470174] env[61985]: DEBUG nova.network.neutron [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Updating instance_info_cache with network_info: [{"id": "6d05ecf8-3c82-464f-be11-a7e2766f9351", "address": "fa:16:3e:eb:4e:cd", "network": {"id": "2f445d2b-5368-4d66-8f25-e10b8eeead9d", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1633536706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e66b7a6e7354b06a77295eadf4e5e0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2c019b6-3ef3-4c8f-95bd-edede2c554a9", "external-id": "nsx-vlan-transportzone-364", "segmentation_id": 364, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d05ecf8-3c", "ovs_interfaceid": "6d05ecf8-3c82-464f-be11-a7e2766f9351", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.795406] env[61985]: INFO nova.compute.manager [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Took 24.59 seconds to build instance. [ 1031.813095] env[61985]: DEBUG oslo_vmware.api [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936380, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.821982] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936381, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.918684] env[61985]: DEBUG nova.network.neutron [-] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.963935] env[61985]: DEBUG nova.compute.utils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1031.968093] env[61985]: DEBUG nova.compute.manager [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1031.968304] env[61985]: DEBUG nova.network.neutron [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1031.975488] env[61985]: DEBUG oslo_concurrency.lockutils [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] Releasing lock "refresh_cache-5cd5e044-b0da-4564-8c2c-e894eb29a74c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.975488] env[61985]: DEBUG nova.compute.manager [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received event network-vif-plugged-0822504d-b73d-40b5-a7a2-4770f9d2f190 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1031.976053] env[61985]: DEBUG oslo_concurrency.lockutils [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] Acquiring lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.976053] env[61985]: DEBUG oslo_concurrency.lockutils [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] Lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.976053] env[61985]: DEBUG oslo_concurrency.lockutils [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] Lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.976248] env[61985]: DEBUG nova.compute.manager [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] No waiting events found dispatching network-vif-plugged-0822504d-b73d-40b5-a7a2-4770f9d2f190 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.976376] env[61985]: WARNING nova.compute.manager [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received unexpected event network-vif-plugged-0822504d-b73d-40b5-a7a2-4770f9d2f190 for instance with vm_state active and task_state None. [ 1031.977170] env[61985]: DEBUG nova.compute.manager [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received event network-changed-0822504d-b73d-40b5-a7a2-4770f9d2f190 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1031.977170] env[61985]: DEBUG nova.compute.manager [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing instance network info cache due to event network-changed-0822504d-b73d-40b5-a7a2-4770f9d2f190. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1031.977170] env[61985]: DEBUG oslo_concurrency.lockutils [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] Acquiring lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.977170] env[61985]: DEBUG oslo_concurrency.lockutils [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] Acquired lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.978969] env[61985]: DEBUG nova.network.neutron [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing network info cache for port 0822504d-b73d-40b5-a7a2-4770f9d2f190 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1032.030219] env[61985]: DEBUG nova.policy [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '344f7f484f4240cd9dc32b200b697e12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfc57e6393c94545be32165d41230db3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1032.298741] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d35961d-fcf3-4868-a6d6-1a1384d837e6 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.098s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.310413] env[61985]: DEBUG oslo_vmware.api [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936380, 'name': ReconfigVM_Task, 'duration_secs': 0.918299} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.311475] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.311475] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Reconfigured VM to attach interface {{(pid=61985) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1032.332418] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.726698} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.332747] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 5cd5e044-b0da-4564-8c2c-e894eb29a74c/5cd5e044-b0da-4564-8c2c-e894eb29a74c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1032.333967] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.335108] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd539320-70e1-4045-8d69-20c7b7c61ed7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.344352] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1032.344352] env[61985]: value = "task-936382" [ 1032.344352] env[61985]: _type = "Task" [ 1032.344352] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.360627] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936382, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.423724] env[61985]: INFO nova.compute.manager [-] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Took 2.23 seconds to deallocate network for instance. [ 1032.468511] env[61985]: DEBUG nova.compute.manager [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1032.473396] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e27d28a8-7015-4e03-b6b0-dad9107f5fbc tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.476017] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.450s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.476017] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.477013] env[61985]: DEBUG oslo_concurrency.lockutils [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.386s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.477236] env[61985]: DEBUG nova.objects.instance [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lazy-loading 'resources' on Instance uuid fea0cf39-e851-409f-86f5-31cc128a44dc {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.507820] env[61985]: INFO nova.scheduler.client.report [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted allocations for instance 625a3143-d138-4b52-aeb7-9e365c6f1128 [ 1032.619339] env[61985]: DEBUG nova.network.neutron [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Successfully created port: 88467e26-c1c7-4b6a-ab2b-9d86b552c0c1 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1032.770985] env[61985]: DEBUG nova.compute.manager [req-172a041c-6faa-4b72-9b82-f1b801c6596a req-1094d37c-23f5-4263-a90b-ad641556bcf8 service nova] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Received event network-vif-deleted-4d18bb14-4f82-46b7-b128-9acc3eaae43f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1032.780148] env[61985]: DEBUG nova.network.neutron [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updated VIF entry in instance network info cache for port 0822504d-b73d-40b5-a7a2-4770f9d2f190. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1032.780584] env[61985]: DEBUG nova.network.neutron [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [{"id": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "address": "fa:16:3e:4b:3d:56", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7075efce-17", "ovs_interfaceid": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0822504d-b73d-40b5-a7a2-4770f9d2f190", "address": "fa:16:3e:ab:c2:ce", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0822504d-b7", "ovs_interfaceid": "0822504d-b73d-40b5-a7a2-4770f9d2f190", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.827299] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bdbb0304-8ee1-4da0-a1e6-e5f42c08cfc6 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5f2aa808-8d1e-471e-9ef7-0f91590ec546-0822504d-b73d-40b5-a7a2-4770f9d2f190" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.740s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.856059] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936382, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085292} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.856422] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.857634] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c88f097-17d5-4bb6-9796-5bc9acc350f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.887185] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 5cd5e044-b0da-4564-8c2c-e894eb29a74c/5cd5e044-b0da-4564-8c2c-e894eb29a74c.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.888286] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f62dc9c4-c714-40d4-a917-73ffd35ec5ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.913048] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1032.913048] env[61985]: value = "task-936383" [ 1032.913048] env[61985]: _type = "Task" [ 1032.913048] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.922711] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936383, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.932136] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.019665] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a4004c5f-0975-4061-8fa7-dcce0b453e90 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "625a3143-d138-4b52-aeb7-9e365c6f1128" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.383s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.160520] env[61985]: DEBUG nova.compute.manager [req-43e62a08-b023-4409-9ef3-ae02ee8b4ffa req-d685ecc9-1e31-4c67-9481-38f8eb047c7f service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Received event network-changed-a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1033.160737] env[61985]: DEBUG nova.compute.manager [req-43e62a08-b023-4409-9ef3-ae02ee8b4ffa req-d685ecc9-1e31-4c67-9481-38f8eb047c7f service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Refreshing instance network info cache due to event network-changed-a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1033.160954] env[61985]: DEBUG oslo_concurrency.lockutils [req-43e62a08-b023-4409-9ef3-ae02ee8b4ffa req-d685ecc9-1e31-4c67-9481-38f8eb047c7f service nova] Acquiring lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.161176] env[61985]: DEBUG oslo_concurrency.lockutils [req-43e62a08-b023-4409-9ef3-ae02ee8b4ffa req-d685ecc9-1e31-4c67-9481-38f8eb047c7f service nova] Acquired lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.161360] env[61985]: DEBUG nova.network.neutron [req-43e62a08-b023-4409-9ef3-ae02ee8b4ffa req-d685ecc9-1e31-4c67-9481-38f8eb047c7f service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Refreshing network info cache for port a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1033.244629] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a65fe51-cf77-4632-8a1a-b52de320872c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.252840] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bdf62a-6a7c-4dc7-9796-4c5d54d72260 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.283973] env[61985]: DEBUG oslo_concurrency.lockutils [req-fb91d8d2-10e9-4f31-af64-c9592f1e19ec req-028ae96b-8571-45c2-a56b-d2fb9605497e service nova] Releasing lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.285239] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ad9ecc-f3eb-444d-90f6-258dcd17d224 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.293293] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d86a39-5b18-413f-8843-2cd274ebf136 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.307890] env[61985]: DEBUG nova.compute.provider_tree [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.424375] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936383, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.479110] env[61985]: DEBUG nova.compute.manager [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1033.506063] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.506374] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.506553] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.506916] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.507155] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.507462] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.507579] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.508682] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.508682] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.508682] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.508682] env[61985]: DEBUG nova.virt.hardware [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.509568] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb5293f-92c6-4856-aa0e-96970c3c9e4a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.518833] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae1d4bd-b809-4f70-9882-5625a462a3a9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.811572] env[61985]: DEBUG nova.scheduler.client.report [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1033.923444] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936383, 'name': ReconfigVM_Task, 'duration_secs': 0.896164} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.923782] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 5cd5e044-b0da-4564-8c2c-e894eb29a74c/5cd5e044-b0da-4564-8c2c-e894eb29a74c.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.924467] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fd3ecbc6-8244-426b-b523-fa22bfdd21cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.931829] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1033.931829] env[61985]: value = "task-936384" [ 1033.931829] env[61985]: _type = "Task" [ 1033.931829] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.940385] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936384, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.132512] env[61985]: DEBUG nova.network.neutron [req-43e62a08-b023-4409-9ef3-ae02ee8b4ffa req-d685ecc9-1e31-4c67-9481-38f8eb047c7f service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updated VIF entry in instance network info cache for port a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.132896] env[61985]: DEBUG nova.network.neutron [req-43e62a08-b023-4409-9ef3-ae02ee8b4ffa req-d685ecc9-1e31-4c67-9481-38f8eb047c7f service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance_info_cache with network_info: [{"id": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "address": "fa:16:3e:e1:d0:1f", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64c525f-b3", "ovs_interfaceid": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.319628] env[61985]: DEBUG oslo_concurrency.lockutils [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.842s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.322322] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.069s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.323962] env[61985]: INFO nova.compute.claims [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1034.349518] env[61985]: INFO nova.scheduler.client.report [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Deleted allocations for instance fea0cf39-e851-409f-86f5-31cc128a44dc [ 1034.387194] env[61985]: DEBUG nova.network.neutron [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Successfully updated port: 88467e26-c1c7-4b6a-ab2b-9d86b552c0c1 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1034.444719] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936384, 'name': Rename_Task, 'duration_secs': 0.351153} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.445050] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1034.445309] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97967d58-0e76-4929-a767-8ccaa7c3468b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.451784] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1034.451784] env[61985]: value = "task-936385" [ 1034.451784] env[61985]: _type = "Task" [ 1034.451784] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.462254] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936385, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.636259] env[61985]: DEBUG oslo_concurrency.lockutils [req-43e62a08-b023-4409-9ef3-ae02ee8b4ffa req-d685ecc9-1e31-4c67-9481-38f8eb047c7f service nova] Releasing lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.846606] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "interface-5f2aa808-8d1e-471e-9ef7-0f91590ec546-0822504d-b73d-40b5-a7a2-4770f9d2f190" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.846828] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5f2aa808-8d1e-471e-9ef7-0f91590ec546-0822504d-b73d-40b5-a7a2-4770f9d2f190" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.859835] env[61985]: DEBUG oslo_concurrency.lockutils [None req-af28f208-03db-4996-bb46-ee591b347b63 tempest-MigrationsAdminTest-701653529 tempest-MigrationsAdminTest-701653529-project-member] Lock "fea0cf39-e851-409f-86f5-31cc128a44dc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.062s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.889800] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.889987] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.890113] env[61985]: DEBUG nova.network.neutron [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1034.962581] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936385, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.197431] env[61985]: DEBUG nova.compute.manager [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Received event network-vif-plugged-88467e26-c1c7-4b6a-ab2b-9d86b552c0c1 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1035.197719] env[61985]: DEBUG oslo_concurrency.lockutils [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] Acquiring lock "8c4a1fe4-4ded-4064-8261-24826181a68c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.197923] env[61985]: DEBUG oslo_concurrency.lockutils [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.198138] env[61985]: DEBUG oslo_concurrency.lockutils [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.198323] env[61985]: DEBUG nova.compute.manager [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] No waiting events found dispatching network-vif-plugged-88467e26-c1c7-4b6a-ab2b-9d86b552c0c1 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1035.198498] env[61985]: WARNING nova.compute.manager [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Received unexpected event network-vif-plugged-88467e26-c1c7-4b6a-ab2b-9d86b552c0c1 for instance with vm_state building and task_state spawning. [ 1035.198691] env[61985]: DEBUG nova.compute.manager [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Received event network-changed-88467e26-c1c7-4b6a-ab2b-9d86b552c0c1 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1035.198872] env[61985]: DEBUG nova.compute.manager [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Refreshing instance network info cache due to event network-changed-88467e26-c1c7-4b6a-ab2b-9d86b552c0c1. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1035.199055] env[61985]: DEBUG oslo_concurrency.lockutils [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] Acquiring lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.349750] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.350068] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.351017] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3691f62f-239c-4fe4-811d-6732fa2dca4c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.377382] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798ab3b9-0599-42aa-ae5f-4cc7c5bc5885 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.406920] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Reconfiguring VM to detach interface {{(pid=61985) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1035.411640] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b9b7891-bae7-40fa-b216-f33fe7e15327 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.433415] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1035.433415] env[61985]: value = "task-936386" [ 1035.433415] env[61985]: _type = "Task" [ 1035.433415] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.445204] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.461956] env[61985]: DEBUG nova.network.neutron [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1035.471066] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936385, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.622832] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9023480-3292-48e2-be93-211df3c794ea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.630610] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a174811-410c-4683-9fd2-6214f3615f8d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.665742] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b638fa-24d9-4f84-be63-4a8b10f933a5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.674241] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82685de-7ea9-4318-9a78-c98e1c43886b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.691191] env[61985]: DEBUG nova.compute.provider_tree [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.713143] env[61985]: DEBUG nova.network.neutron [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance_info_cache with network_info: [{"id": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "address": "fa:16:3e:35:b3:a2", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88467e26-c1", "ovs_interfaceid": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.943647] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.962248] env[61985]: DEBUG oslo_vmware.api [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936385, 'name': PowerOnVM_Task, 'duration_secs': 1.033851} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.962605] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1035.962835] env[61985]: INFO nova.compute.manager [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Took 9.01 seconds to spawn the instance on the hypervisor. [ 1035.963075] env[61985]: DEBUG nova.compute.manager [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.964021] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ab545c-8a9e-415a-9fee-f87fc1d5161c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.194378] env[61985]: DEBUG nova.scheduler.client.report [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.216371] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.216582] env[61985]: DEBUG nova.compute.manager [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Instance network_info: |[{"id": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "address": "fa:16:3e:35:b3:a2", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88467e26-c1", "ovs_interfaceid": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1036.217045] env[61985]: DEBUG oslo_concurrency.lockutils [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] Acquired lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.217111] env[61985]: DEBUG nova.network.neutron [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Refreshing network info cache for port 88467e26-c1c7-4b6a-ab2b-9d86b552c0c1 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1036.219523] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:b3:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bff6c3a1-cc80-46ca-86c0-6dbb029edddb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88467e26-c1c7-4b6a-ab2b-9d86b552c0c1', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1036.225703] env[61985]: DEBUG oslo.service.loopingcall [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.226369] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1036.226612] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ddf93ccd-5a20-4028-acd4-47e3cd825bab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.247493] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1036.247493] env[61985]: value = "task-936387" [ 1036.247493] env[61985]: _type = "Task" [ 1036.247493] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.255692] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936387, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.443333] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.481833] env[61985]: INFO nova.compute.manager [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Took 23.22 seconds to build instance. [ 1036.700174] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.700757] env[61985]: DEBUG nova.compute.manager [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1036.703761] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.910s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.703998] env[61985]: DEBUG nova.objects.instance [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lazy-loading 'resources' on Instance uuid afa46da0-3bd1-47ea-bdf0-28eed87fd627 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.758815] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936387, 'name': CreateVM_Task, 'duration_secs': 0.364923} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.759068] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1036.759864] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.760243] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.760518] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1036.761117] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af47d4a4-68d6-4a88-8faa-58c3c1d275a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.766690] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1036.766690] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5209cef2-11ed-b5c8-98a0-7da4f7fb72f5" [ 1036.766690] env[61985]: _type = "Task" [ 1036.766690] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.779458] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5209cef2-11ed-b5c8-98a0-7da4f7fb72f5, 'name': SearchDatastore_Task, 'duration_secs': 0.009955} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.779839] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.780195] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1036.780503] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.780735] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.780990] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1036.781341] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-784c1aef-863a-434d-a99a-69c91573048b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.790405] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1036.790672] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1036.791532] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f7bb6d7-5270-44e1-b42e-4ab4f6734848 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.797949] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1036.797949] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d096b3-424e-6ecc-554f-ead864a373d9" [ 1036.797949] env[61985]: _type = "Task" [ 1036.797949] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.808823] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d096b3-424e-6ecc-554f-ead864a373d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.945073] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.984721] env[61985]: DEBUG nova.network.neutron [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updated VIF entry in instance network info cache for port 88467e26-c1c7-4b6a-ab2b-9d86b552c0c1. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1036.985134] env[61985]: DEBUG nova.network.neutron [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance_info_cache with network_info: [{"id": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "address": "fa:16:3e:35:b3:a2", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88467e26-c1", "ovs_interfaceid": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.986577] env[61985]: DEBUG oslo_concurrency.lockutils [None req-59bd7dde-6d41-4870-8d58-922452c9c8d0 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.119s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.206685] env[61985]: DEBUG nova.objects.instance [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lazy-loading 'numa_topology' on Instance uuid afa46da0-3bd1-47ea-bdf0-28eed87fd627 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.211019] env[61985]: DEBUG nova.compute.utils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1037.211019] env[61985]: DEBUG nova.compute.manager [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1037.211019] env[61985]: DEBUG nova.network.neutron [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1037.259675] env[61985]: DEBUG nova.policy [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e23fdb33638143409613d9d4f26f199c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe531ec9f60548adb457351e19f8d51b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1037.309697] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d096b3-424e-6ecc-554f-ead864a373d9, 'name': SearchDatastore_Task, 'duration_secs': 0.011746} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.310528] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-531298c8-72f6-4a04-8913-38002d611602 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.316058] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1037.316058] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52239fbb-1252-4728-8242-d28dd4418edc" [ 1037.316058] env[61985]: _type = "Task" [ 1037.316058] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.325827] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52239fbb-1252-4728-8242-d28dd4418edc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.446122] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.490309] env[61985]: DEBUG oslo_concurrency.lockutils [req-878f8d8f-1faf-416e-ba02-66621b53122f req-ee16a6b2-6e78-4bed-8f66-0f4ba088c8d3 service nova] Releasing lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.519292] env[61985]: DEBUG nova.network.neutron [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Successfully created port: 36a3c11a-0e39-4b68-ada2-17d3d6a82a56 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.710886] env[61985]: DEBUG nova.objects.base [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1037.715164] env[61985]: DEBUG nova.compute.manager [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1037.831755] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52239fbb-1252-4728-8242-d28dd4418edc, 'name': SearchDatastore_Task, 'duration_secs': 0.009389} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.832053] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.832365] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 8c4a1fe4-4ded-4064-8261-24826181a68c/8c4a1fe4-4ded-4064-8261-24826181a68c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1037.832596] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5d6a4eaa-5964-40b7-9733-cbcaae2eb4cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.842546] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1037.842546] env[61985]: value = "task-936388" [ 1037.842546] env[61985]: _type = "Task" [ 1037.842546] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.852259] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936388, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.884495] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527e3c7d-abd8-2dd4-bb92-6f207d0d7d26/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1037.885526] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c802fee-f3e5-4ee5-ae86-be3e524262c8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.894229] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527e3c7d-abd8-2dd4-bb92-6f207d0d7d26/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1037.894419] env[61985]: ERROR oslo_vmware.rw_handles [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527e3c7d-abd8-2dd4-bb92-6f207d0d7d26/disk-0.vmdk due to incomplete transfer. [ 1037.894660] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1e46ec2e-ee98-4890-bfa0-1f07dc4f7d4d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.902685] env[61985]: DEBUG oslo_vmware.rw_handles [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527e3c7d-abd8-2dd4-bb92-6f207d0d7d26/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1037.902892] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Uploaded image 173db284-6b8b-4355-bdde-6a9ba3be888e to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1037.905281] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1037.905608] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-21349aca-5814-4e9e-beac-36b129d43b80 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.911350] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1037.911350] env[61985]: value = "task-936389" [ 1037.911350] env[61985]: _type = "Task" [ 1037.911350] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.923414] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936389, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.945261] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.984134] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc61b49b-23d6-40f6-9c52-6e4cd9d48de1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.991929] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e229ce-6c69-47c9-a68d-dd2e25fd0ef2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.026055] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae33209-056f-418b-8d1c-0fc69e726a9b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.034223] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bea64c-a271-46a6-bfaf-6c5db1c99d75 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.049487] env[61985]: DEBUG nova.compute.provider_tree [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.222138] env[61985]: DEBUG nova.compute.manager [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.227022] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9e530d-84a0-44a7-b536-24f9c0330d50 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.353492] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936388, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497645} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.353784] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 8c4a1fe4-4ded-4064-8261-24826181a68c/8c4a1fe4-4ded-4064-8261-24826181a68c.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1038.354029] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1038.354293] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-187c32c3-249c-41b1-8399-9492a13df124 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.361186] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1038.361186] env[61985]: value = "task-936390" [ 1038.361186] env[61985]: _type = "Task" [ 1038.361186] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.368997] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936390, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.422051] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936389, 'name': Destroy_Task, 'duration_secs': 0.497139} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.422051] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Destroyed the VM [ 1038.422051] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1038.422051] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c5ebb3cd-61bd-48f3-849a-3f1e09e58b61 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.428314] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1038.428314] env[61985]: value = "task-936391" [ 1038.428314] env[61985]: _type = "Task" [ 1038.428314] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.439046] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936391, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.455779] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.552608] env[61985]: DEBUG nova.scheduler.client.report [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.739094] env[61985]: DEBUG nova.compute.manager [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1038.742655] env[61985]: INFO nova.compute.manager [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] instance snapshotting [ 1038.751655] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68eb7637-a072-4006-a82c-12b9c0d8376e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.778147] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759dbe98-b1de-44f5-9e10-fbac22b382e8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.792565] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.792935] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.793307] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.793307] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.793449] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.793487] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.793818] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.794062] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.794217] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.794392] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.795236] env[61985]: DEBUG nova.virt.hardware [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.796605] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74c55c9-82df-4228-8258-4afcfaea594f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.805406] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f904ab1-1c75-4be3-ad63-357fbac65e8f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.872657] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936390, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.12112} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.872657] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1038.872833] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7ac1ce-4ce7-4c8a-a2a5-1351ca4e1412 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.898322] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 8c4a1fe4-4ded-4064-8261-24826181a68c/8c4a1fe4-4ded-4064-8261-24826181a68c.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.898720] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e771b1f5-75e5-4b96-886c-dc5a94ff0f99 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.922645] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1038.922645] env[61985]: value = "task-936392" [ 1038.922645] env[61985]: _type = "Task" [ 1038.922645] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.929045] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936392, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.937401] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936391, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.948170] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.058956] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.355s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.066373] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.598s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.066373] env[61985]: DEBUG nova.objects.instance [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lazy-loading 'resources' on Instance uuid 49707589-4969-4e08-882b-2a2c94bc0d85 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.135254] env[61985]: DEBUG nova.compute.manager [req-4946cee9-39d6-469d-872e-ef742ca827eb req-5be029a5-3da0-4053-9483-4a4feb6d4866 service nova] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Received event network-vif-plugged-36a3c11a-0e39-4b68-ada2-17d3d6a82a56 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1039.135458] env[61985]: DEBUG oslo_concurrency.lockutils [req-4946cee9-39d6-469d-872e-ef742ca827eb req-5be029a5-3da0-4053-9483-4a4feb6d4866 service nova] Acquiring lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.135527] env[61985]: DEBUG oslo_concurrency.lockutils [req-4946cee9-39d6-469d-872e-ef742ca827eb req-5be029a5-3da0-4053-9483-4a4feb6d4866 service nova] Lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.135715] env[61985]: DEBUG oslo_concurrency.lockutils [req-4946cee9-39d6-469d-872e-ef742ca827eb req-5be029a5-3da0-4053-9483-4a4feb6d4866 service nova] Lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.135894] env[61985]: DEBUG nova.compute.manager [req-4946cee9-39d6-469d-872e-ef742ca827eb req-5be029a5-3da0-4053-9483-4a4feb6d4866 service nova] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] No waiting events found dispatching network-vif-plugged-36a3c11a-0e39-4b68-ada2-17d3d6a82a56 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1039.137191] env[61985]: WARNING nova.compute.manager [req-4946cee9-39d6-469d-872e-ef742ca827eb req-5be029a5-3da0-4053-9483-4a4feb6d4866 service nova] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Received unexpected event network-vif-plugged-36a3c11a-0e39-4b68-ada2-17d3d6a82a56 for instance with vm_state building and task_state spawning. [ 1039.146846] env[61985]: DEBUG nova.network.neutron [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Successfully updated port: 36a3c11a-0e39-4b68-ada2-17d3d6a82a56 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1039.301347] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1039.301668] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ed18db05-7283-47a2-abb8-527cfa18e63f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.311507] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1039.311507] env[61985]: value = "task-936393" [ 1039.311507] env[61985]: _type = "Task" [ 1039.311507] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.320089] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936393, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.436198] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936392, 'name': ReconfigVM_Task, 'duration_secs': 0.291593} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.436859] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 8c4a1fe4-4ded-4064-8261-24826181a68c/8c4a1fe4-4ded-4064-8261-24826181a68c.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.437502] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f602ed11-10a3-413c-92ec-83dfb8acc371 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.441763] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936391, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.450427] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.450679] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1039.450679] env[61985]: value = "task-936394" [ 1039.450679] env[61985]: _type = "Task" [ 1039.450679] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.458688] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936394, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.577836] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2aac4036-77db-470d-9e6d-e5b998c9494a tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 38.362s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.581839] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 17.429s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.581839] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.581839] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.581839] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.581839] env[61985]: INFO nova.compute.manager [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Terminating instance [ 1039.582971] env[61985]: DEBUG nova.compute.manager [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1039.583217] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1039.583450] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-696744f4-e0e4-4fb7-b9ca-5b9d6b299d8b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.594262] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64cf3c18-9100-4053-9bb5-f9aa031942ca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.631263] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance afa46da0-3bd1-47ea-bdf0-28eed87fd627 could not be found. [ 1039.631507] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1039.631699] env[61985]: INFO nova.compute.manager [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1039.631961] env[61985]: DEBUG oslo.service.loopingcall [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.634727] env[61985]: DEBUG nova.compute.manager [-] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.634867] env[61985]: DEBUG nova.network.neutron [-] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1039.652399] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "refresh_cache-5aedbb9e-8bd1-499a-8195-b3045f7a141f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.652536] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "refresh_cache-5aedbb9e-8bd1-499a-8195-b3045f7a141f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.652681] env[61985]: DEBUG nova.network.neutron [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1039.829182] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936393, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.858985] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31127fd-cf43-4fd6-b620-2c1120caccbe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.866726] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0535d24a-3109-4a00-b3a2-399ba0e5ec56 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.903245] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1febf722-e040-4b50-a6c1-5b23a8fa4e2c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.913538] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2b08f0-a9f5-42e2-a15a-b9401ab536af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.929432] env[61985]: DEBUG nova.compute.provider_tree [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.938795] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936391, 'name': RemoveSnapshot_Task, 'duration_secs': 1.160815} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.942269] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1039.942269] env[61985]: DEBUG nova.compute.manager [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1039.942269] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0858ef0f-79cb-49f6-a91e-70221f2681d9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.958097] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.966476] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936394, 'name': Rename_Task, 'duration_secs': 0.137304} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.966749] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1039.967054] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7754183-e189-4772-a865-eb75b4d8e2cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.973345] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1039.973345] env[61985]: value = "task-936395" [ 1039.973345] env[61985]: _type = "Task" [ 1039.973345] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.981505] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936395, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.184160] env[61985]: DEBUG nova.network.neutron [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1040.319479] env[61985]: DEBUG nova.network.neutron [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Updating instance_info_cache with network_info: [{"id": "36a3c11a-0e39-4b68-ada2-17d3d6a82a56", "address": "fa:16:3e:a1:51:a4", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36a3c11a-0e", "ovs_interfaceid": "36a3c11a-0e39-4b68-ada2-17d3d6a82a56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.328669] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936393, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.435522] env[61985]: DEBUG nova.scheduler.client.report [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.454277] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.461990] env[61985]: INFO nova.compute.manager [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Shelve offloading [ 1040.463705] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1040.463955] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c81ddac-c5b2-41cb-b563-6bf0d3d450d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.470032] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1040.470032] env[61985]: value = "task-936396" [ 1040.470032] env[61985]: _type = "Task" [ 1040.470032] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.487210] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936395, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.488048] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1040.488346] env[61985]: DEBUG nova.compute.manager [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1040.489199] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df32565e-560b-40a4-9981-1ae3f04f98d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.494872] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.495129] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.495412] env[61985]: DEBUG nova.network.neutron [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1040.649986] env[61985]: DEBUG nova.network.neutron [-] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.823706] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936393, 'name': CreateSnapshot_Task, 'duration_secs': 1.294324} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.824089] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1040.825178] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5ce001-5ac6-42af-8e7c-d0edf5b76219 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.828069] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "refresh_cache-5aedbb9e-8bd1-499a-8195-b3045f7a141f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.830326] env[61985]: DEBUG nova.compute.manager [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Instance network_info: |[{"id": "36a3c11a-0e39-4b68-ada2-17d3d6a82a56", "address": "fa:16:3e:a1:51:a4", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36a3c11a-0e", "ovs_interfaceid": "36a3c11a-0e39-4b68-ada2-17d3d6a82a56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1040.830326] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:51:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55520f67-d092-4eb7-940f-d7cceaa1ca1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '36a3c11a-0e39-4b68-ada2-17d3d6a82a56', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1040.836606] env[61985]: DEBUG oslo.service.loopingcall [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.837436] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1040.837724] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae57730b-1ac2-49fb-9ad7-ebb17b3e77f4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.864821] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1040.864821] env[61985]: value = "task-936397" [ 1040.864821] env[61985]: _type = "Task" [ 1040.864821] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.871860] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936397, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.944018] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.877s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.944018] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.164s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.945391] env[61985]: DEBUG nova.objects.instance [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lazy-loading 'resources' on Instance uuid 90ac7d3e-aef3-4a50-963a-606e7c58f446 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.958150] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.967287] env[61985]: INFO nova.scheduler.client.report [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted allocations for instance 49707589-4969-4e08-882b-2a2c94bc0d85 [ 1040.984732] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936395, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.152712] env[61985]: INFO nova.compute.manager [-] [instance: afa46da0-3bd1-47ea-bdf0-28eed87fd627] Took 1.52 seconds to deallocate network for instance. [ 1041.188587] env[61985]: DEBUG nova.compute.manager [req-c5b173a3-6987-467d-8651-2ba3f2faea14 req-07b85f10-fceb-4a37-aaf5-1c33028f332a service nova] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Received event network-changed-36a3c11a-0e39-4b68-ada2-17d3d6a82a56 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1041.188822] env[61985]: DEBUG nova.compute.manager [req-c5b173a3-6987-467d-8651-2ba3f2faea14 req-07b85f10-fceb-4a37-aaf5-1c33028f332a service nova] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Refreshing instance network info cache due to event network-changed-36a3c11a-0e39-4b68-ada2-17d3d6a82a56. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1041.190134] env[61985]: DEBUG oslo_concurrency.lockutils [req-c5b173a3-6987-467d-8651-2ba3f2faea14 req-07b85f10-fceb-4a37-aaf5-1c33028f332a service nova] Acquiring lock "refresh_cache-5aedbb9e-8bd1-499a-8195-b3045f7a141f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.190364] env[61985]: DEBUG oslo_concurrency.lockutils [req-c5b173a3-6987-467d-8651-2ba3f2faea14 req-07b85f10-fceb-4a37-aaf5-1c33028f332a service nova] Acquired lock "refresh_cache-5aedbb9e-8bd1-499a-8195-b3045f7a141f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.190583] env[61985]: DEBUG nova.network.neutron [req-c5b173a3-6987-467d-8651-2ba3f2faea14 req-07b85f10-fceb-4a37-aaf5-1c33028f332a service nova] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Refreshing network info cache for port 36a3c11a-0e39-4b68-ada2-17d3d6a82a56 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1041.368883] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1041.370117] env[61985]: DEBUG nova.network.neutron [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updating instance_info_cache with network_info: [{"id": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "address": "fa:16:3e:af:c8:fd", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47c6b710-95", "ovs_interfaceid": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.371401] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-58b59e26-2e77-4ff7-af62-186d4732fbea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.387248] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936397, 'name': CreateVM_Task, 'duration_secs': 0.512204} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.388567] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1041.389020] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1041.389020] env[61985]: value = "task-936398" [ 1041.389020] env[61985]: _type = "Task" [ 1041.389020] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.389628] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.389799] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.390140] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1041.390701] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e681fd9-4f01-4b69-b84a-4d49a0b89dc8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.399623] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1041.399623] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52819c2a-498d-8373-aee8-cc891ab24857" [ 1041.399623] env[61985]: _type = "Task" [ 1041.399623] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.402812] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936398, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.411643] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52819c2a-498d-8373-aee8-cc891ab24857, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.464249] env[61985]: DEBUG oslo_vmware.api [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936386, 'name': ReconfigVM_Task, 'duration_secs': 5.803249} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.464249] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.464249] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Reconfigured VM to detach interface {{(pid=61985) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1041.479232] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb9c4544-f272-48de-be3a-85f314dc85b4 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "49707589-4969-4e08-882b-2a2c94bc0d85" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.433s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.492546] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936395, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.729376] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ab1f13-41a5-4625-a697-c5fc9563af22 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.740055] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8b335d-8b29-4ec8-bff1-83aee614c93d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.784762] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18a21f7-e07c-47b9-bc69-75f43c2b8a83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.792908] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b96dc9f-46a7-4eaf-92d1-b58f3710c280 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.808747] env[61985]: DEBUG nova.compute.provider_tree [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.878215] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.905336] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936398, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.914469] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52819c2a-498d-8373-aee8-cc891ab24857, 'name': SearchDatastore_Task, 'duration_secs': 0.014637} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.914792] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.915160] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1041.915457] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.915647] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.915885] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1041.916135] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d98c821-d397-49bc-9708-470dd1c53c00 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.924304] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1041.924496] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1041.925237] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61a4e991-b7e5-4baa-9e6c-4174fd114962 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.931479] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1041.931479] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52530d46-99ea-cb1c-4463-fb0eee2827a7" [ 1041.931479] env[61985]: _type = "Task" [ 1041.931479] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.940448] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52530d46-99ea-cb1c-4463-fb0eee2827a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.989710] env[61985]: DEBUG oslo_vmware.api [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936395, 'name': PowerOnVM_Task, 'duration_secs': 1.567855} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.989993] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1041.990407] env[61985]: INFO nova.compute.manager [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Took 8.51 seconds to spawn the instance on the hypervisor. [ 1041.990630] env[61985]: DEBUG nova.compute.manager [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1041.992773] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ba7a9b-1a50-4670-b527-f2f41e3ee008 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.148834] env[61985]: DEBUG nova.network.neutron [req-c5b173a3-6987-467d-8651-2ba3f2faea14 req-07b85f10-fceb-4a37-aaf5-1c33028f332a service nova] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Updated VIF entry in instance network info cache for port 36a3c11a-0e39-4b68-ada2-17d3d6a82a56. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1042.149303] env[61985]: DEBUG nova.network.neutron [req-c5b173a3-6987-467d-8651-2ba3f2faea14 req-07b85f10-fceb-4a37-aaf5-1c33028f332a service nova] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Updating instance_info_cache with network_info: [{"id": "36a3c11a-0e39-4b68-ada2-17d3d6a82a56", "address": "fa:16:3e:a1:51:a4", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36a3c11a-0e", "ovs_interfaceid": "36a3c11a-0e39-4b68-ada2-17d3d6a82a56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.186666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5953357f-d6ec-48ea-976f-f2b459f9ba9d tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "afa46da0-3bd1-47ea-bdf0-28eed87fd627" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.608s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.251186] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1042.251186] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac4ac55-adda-42f6-a880-3b15122b688a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.258335] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1042.258821] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8c28cfe0-5fca-43e5-aede-e37e80f7bb93 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.313176] env[61985]: DEBUG nova.scheduler.client.report [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.322348] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1042.322348] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1042.322348] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleting the datastore file [datastore1] 87006fbf-ea90-4d9a-88af-001de424ac14 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.322348] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9cdde035-6759-4127-9d2f-1d1b6b70bbc4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.328510] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1042.328510] env[61985]: value = "task-936400" [ 1042.328510] env[61985]: _type = "Task" [ 1042.328510] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.340188] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.405320] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936398, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.442805] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52530d46-99ea-cb1c-4463-fb0eee2827a7, 'name': SearchDatastore_Task, 'duration_secs': 0.009873} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.444313] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1533e93c-5c26-4279-8370-742bbe09b141 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.451055] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1042.451055] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5216d2ae-ca25-892c-0df1-1ef2464022dd" [ 1042.451055] env[61985]: _type = "Task" [ 1042.451055] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.460160] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5216d2ae-ca25-892c-0df1-1ef2464022dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.514997] env[61985]: INFO nova.compute.manager [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Took 24.14 seconds to build instance. [ 1042.652322] env[61985]: DEBUG oslo_concurrency.lockutils [req-c5b173a3-6987-467d-8651-2ba3f2faea14 req-07b85f10-fceb-4a37-aaf5-1c33028f332a service nova] Releasing lock "refresh_cache-5aedbb9e-8bd1-499a-8195-b3045f7a141f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.820237] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.874s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.823189] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.899s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.823942] env[61985]: DEBUG nova.objects.instance [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lazy-loading 'resources' on Instance uuid cfd59b61-cca9-48d5-85e1-1f45d13f1e88 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.842803] env[61985]: DEBUG oslo_vmware.api [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167087} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.843319] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.843622] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1042.844020] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1042.858667] env[61985]: INFO nova.scheduler.client.report [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Deleted allocations for instance 90ac7d3e-aef3-4a50-963a-606e7c58f446 [ 1042.897695] env[61985]: INFO nova.scheduler.client.report [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleted allocations for instance 87006fbf-ea90-4d9a-88af-001de424ac14 [ 1042.910517] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936398, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.962699] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5216d2ae-ca25-892c-0df1-1ef2464022dd, 'name': SearchDatastore_Task, 'duration_secs': 0.009785} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.963167] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.963721] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 5aedbb9e-8bd1-499a-8195-b3045f7a141f/5aedbb9e-8bd1-499a-8195-b3045f7a141f.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1042.964171] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae7546ef-1017-4fdf-8f7b-37b693c07eb1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.974341] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1042.974341] env[61985]: value = "task-936401" [ 1042.974341] env[61985]: _type = "Task" [ 1042.974341] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.983825] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936401, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.004548] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.004756] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.004929] env[61985]: DEBUG nova.network.neutron [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1043.016929] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a64f67be-e8eb-45ab-8790-f929a97f5a59 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.651s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.316012] env[61985]: DEBUG nova.compute.manager [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Received event network-vif-unplugged-47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1043.316264] env[61985]: DEBUG oslo_concurrency.lockutils [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] Acquiring lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.316528] env[61985]: DEBUG oslo_concurrency.lockutils [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] Lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.316657] env[61985]: DEBUG oslo_concurrency.lockutils [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] Lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.316835] env[61985]: DEBUG nova.compute.manager [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] No waiting events found dispatching network-vif-unplugged-47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1043.317017] env[61985]: WARNING nova.compute.manager [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Received unexpected event network-vif-unplugged-47c6b710-9517-4551-b70b-d11e6cdd7c8f for instance with vm_state shelved_offloaded and task_state None. [ 1043.318634] env[61985]: DEBUG nova.compute.manager [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Received event network-changed-47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1043.318818] env[61985]: DEBUG nova.compute.manager [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Refreshing instance network info cache due to event network-changed-47c6b710-9517-4551-b70b-d11e6cdd7c8f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1043.319091] env[61985]: DEBUG oslo_concurrency.lockutils [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] Acquiring lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.319180] env[61985]: DEBUG oslo_concurrency.lockutils [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] Acquired lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.319344] env[61985]: DEBUG nova.network.neutron [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Refreshing network info cache for port 47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1043.370417] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1dd92a0e-4610-4a2d-8a88-6e5a6f820ece tempest-ImagesOneServerNegativeTestJSON-834074737 tempest-ImagesOneServerNegativeTestJSON-834074737-project-member] Lock "90ac7d3e-aef3-4a50-963a-606e7c58f446" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.798s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.409488] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.415233] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936398, 'name': CloneVM_Task, 'duration_secs': 1.575271} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.418813] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Created linked-clone VM from snapshot [ 1043.419918] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daef6c06-9697-4bfb-8456-82716343181d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.439275] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Uploading image 74ed60d3-a3a4-4ead-b9e2-77397e41a88d {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1043.456627] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1043.456942] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-dbc77e09-e931-49eb-9292-3e0b07b537f6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.467034] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1043.467034] env[61985]: value = "task-936402" [ 1043.467034] env[61985]: _type = "Task" [ 1043.467034] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.484673] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936402, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.491213] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936401, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.506415] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.507028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.595944] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed26d48a-5232-4dca-8d64-9fe6b9167f19 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.610215] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae0a8bf-add6-4c72-85b2-a3850b574f81 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.645337] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b025fa-dff6-45c1-a508-adc64007c893 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.655316] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c86efeb-9749-44f8-a424-d3e57b0c0b0d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.674466] env[61985]: DEBUG nova.compute.provider_tree [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.935529] env[61985]: DEBUG nova.compute.manager [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Stashing vm_state: active {{(pid=61985) _prep_resize /opt/stack/nova/nova/compute/manager.py:5630}} [ 1043.980561] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936402, 'name': Destroy_Task} progress is 33%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.993248] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936401, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.845208} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.993799] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 5aedbb9e-8bd1-499a-8195-b3045f7a141f/5aedbb9e-8bd1-499a-8195-b3045f7a141f.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1043.994153] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1043.994444] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34034c2a-11c0-48a4-ae8a-51c2518a3fbd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.005290] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1044.005290] env[61985]: value = "task-936403" [ 1044.005290] env[61985]: _type = "Task" [ 1044.005290] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.017030] env[61985]: DEBUG nova.compute.manager [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1044.019801] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936403, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.117930] env[61985]: INFO nova.network.neutron [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Port 0822504d-b73d-40b5-a7a2-4770f9d2f190 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1044.118361] env[61985]: DEBUG nova.network.neutron [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [{"id": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "address": "fa:16:3e:4b:3d:56", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7075efce-17", "ovs_interfaceid": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.178954] env[61985]: DEBUG nova.scheduler.client.report [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.263513] env[61985]: DEBUG nova.network.neutron [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updated VIF entry in instance network info cache for port 47c6b710-9517-4551-b70b-d11e6cdd7c8f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1044.263965] env[61985]: DEBUG nova.network.neutron [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updating instance_info_cache with network_info: [{"id": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "address": "fa:16:3e:af:c8:fd", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap47c6b710-95", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.455038] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.480187] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936402, 'name': Destroy_Task, 'duration_secs': 0.557845} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.481598] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Destroyed the VM [ 1044.481967] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1044.482310] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-077b2896-55da-47a4-9bd8-c18f5af065e5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.492748] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1044.492748] env[61985]: value = "task-936404" [ 1044.492748] env[61985]: _type = "Task" [ 1044.492748] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.506151] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936404, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.516823] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936403, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.20922} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.520963] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1044.520963] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37b4c3f-c6bd-4b3a-9f37-3eab84c32315 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.552278] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 5aedbb9e-8bd1-499a-8195-b3045f7a141f/5aedbb9e-8bd1-499a-8195-b3045f7a141f.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1044.553637] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.554394] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52271803-88cd-4ed8-a342-180b74b29e8a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.579677] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1044.579677] env[61985]: value = "task-936405" [ 1044.579677] env[61985]: _type = "Task" [ 1044.579677] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.587500] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936405, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.621442] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.686446] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.864s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.687784] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.755s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.691664] env[61985]: DEBUG nova.objects.instance [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lazy-loading 'resources' on Instance uuid ce757dda-f58a-47b3-b319-e6b01f05b20b {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.707299] env[61985]: INFO nova.scheduler.client.report [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Deleted allocations for instance cfd59b61-cca9-48d5-85e1-1f45d13f1e88 [ 1044.765090] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "036fae78-2af0-49f1-85fa-245c29f35ebc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.765090] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.770888] env[61985]: DEBUG oslo_concurrency.lockutils [req-3727684a-a51a-49f0-a71c-41025def1fb6 req-812a0b0f-8c24-4fa5-ac66-4e040fd001e1 service nova] Releasing lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.011390] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936404, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.092192] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936405, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.127524] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8c78d8c1-5fd7-4aa8-9a26-f3d185951b30 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-5f2aa808-8d1e-471e-9ef7-0f91590ec546-0822504d-b73d-40b5-a7a2-4770f9d2f190" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.279s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.217149] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b83e6370-2a77-45fb-9bec-4ca23a68e7e7 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "cfd59b61-cca9-48d5-85e1-1f45d13f1e88" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.445s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.266926] env[61985]: DEBUG nova.compute.manager [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1045.405387] env[61985]: DEBUG nova.compute.manager [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received event network-changed-7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1045.405599] env[61985]: DEBUG nova.compute.manager [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing instance network info cache due to event network-changed-7075efce-177a-4127-9c2f-7e3a9d1e92ad. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1045.405762] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] Acquiring lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.405903] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] Acquired lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.406080] env[61985]: DEBUG nova.network.neutron [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Refreshing network info cache for port 7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1045.451235] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5082332e-42b3-47e3-b815-df6e06a29e46 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.460586] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d41ec86-3901-4a66-bba9-2a163b64d34e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.507613] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a3ac41-fc52-4882-ab8b-ce05c09c59b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.519838] env[61985]: DEBUG oslo_vmware.api [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936404, 'name': RemoveSnapshot_Task, 'duration_secs': 0.86821} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.521740] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459ee9dd-a233-40d2-9691-462098a1b8ac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.527635] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1045.543937] env[61985]: DEBUG nova.compute.provider_tree [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.589504] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936405, 'name': ReconfigVM_Task, 'duration_secs': 0.640103} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.589789] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 5aedbb9e-8bd1-499a-8195-b3045f7a141f/5aedbb9e-8bd1-499a-8195-b3045f7a141f.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1045.590457] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc5d4250-34e4-43da-b07e-bee18aeab080 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.598287] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1045.598287] env[61985]: value = "task-936406" [ 1045.598287] env[61985]: _type = "Task" [ 1045.598287] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.607511] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936406, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.799064] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.005260] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "87006fbf-ea90-4d9a-88af-001de424ac14" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.031993] env[61985]: WARNING nova.compute.manager [None req-1c6339af-1441-4453-b556-6315565adcca tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Image not found during snapshot: nova.exception.ImageNotFound: Image 74ed60d3-a3a4-4ead-b9e2-77397e41a88d could not be found. [ 1046.046515] env[61985]: DEBUG nova.scheduler.client.report [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.109306] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936406, 'name': Rename_Task, 'duration_secs': 0.308481} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.109559] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1046.110080] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6863134a-a047-4ced-8e36-5768322f566d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.119415] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1046.119415] env[61985]: value = "task-936407" [ 1046.119415] env[61985]: _type = "Task" [ 1046.119415] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.128416] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936407, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.128813] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.129064] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.129283] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.129483] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.129713] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.131793] env[61985]: INFO nova.compute.manager [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Terminating instance [ 1046.136074] env[61985]: DEBUG nova.compute.manager [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1046.136074] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1046.136074] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682b32b6-69c7-4356-b74c-98a8bebf8305 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.142890] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1046.143194] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9adfc3a-6946-40a1-9f11-b35c5e97bc68 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.151906] env[61985]: DEBUG oslo_vmware.api [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 1046.151906] env[61985]: value = "task-936408" [ 1046.151906] env[61985]: _type = "Task" [ 1046.151906] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.162751] env[61985]: DEBUG oslo_vmware.api [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936408, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.343884] env[61985]: DEBUG nova.network.neutron [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updated VIF entry in instance network info cache for port 7075efce-177a-4127-9c2f-7e3a9d1e92ad. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.343884] env[61985]: DEBUG nova.network.neutron [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [{"id": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "address": "fa:16:3e:4b:3d:56", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7075efce-17", "ovs_interfaceid": "7075efce-177a-4127-9c2f-7e3a9d1e92ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.555064] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.868s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.562898] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.150s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.562898] env[61985]: DEBUG nova.objects.instance [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lazy-loading 'resources' on Instance uuid 87006fbf-ea90-4d9a-88af-001de424ac14 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.594758] env[61985]: INFO nova.scheduler.client.report [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleted allocations for instance ce757dda-f58a-47b3-b319-e6b01f05b20b [ 1046.632953] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936407, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.664907] env[61985]: DEBUG oslo_vmware.api [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936408, 'name': PowerOffVM_Task, 'duration_secs': 0.216466} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.665211] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1046.665392] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1046.665677] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68f204c7-751a-4029-939a-4318ff9c7be2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.756454] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1046.756795] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1046.757063] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Deleting the datastore file [datastore2] 761d5954-a5ca-4459-a1d6-bfc59b284bf4 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.758143] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b540fe9f-ed1b-4dfa-aaff-c451ad6efd7b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.771977] env[61985]: DEBUG oslo_vmware.api [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for the task: (returnval){ [ 1046.771977] env[61985]: value = "task-936410" [ 1046.771977] env[61985]: _type = "Task" [ 1046.771977] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.783429] env[61985]: DEBUG oslo_vmware.api [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936410, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.846292] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] Releasing lock "refresh_cache-5f2aa808-8d1e-471e-9ef7-0f91590ec546" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.847574] env[61985]: DEBUG nova.compute.manager [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Received event network-changed-88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1046.847851] env[61985]: DEBUG nova.compute.manager [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Refreshing instance network info cache due to event network-changed-88a63ff6-55f6-4b8b-ba9e-188245f7ea87. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1046.848140] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] Acquiring lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.848307] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] Acquired lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.848484] env[61985]: DEBUG nova.network.neutron [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Refreshing network info cache for port 88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1047.062197] env[61985]: DEBUG nova.objects.instance [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lazy-loading 'numa_topology' on Instance uuid 87006fbf-ea90-4d9a-88af-001de424ac14 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.108183] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ae874afb-a4e9-4786-9445-1c247eeb2f15 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "ce757dda-f58a-47b3-b319-e6b01f05b20b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.023s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.136462] env[61985]: DEBUG oslo_vmware.api [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936407, 'name': PowerOnVM_Task, 'duration_secs': 0.818027} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.136599] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1047.137374] env[61985]: INFO nova.compute.manager [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Took 8.40 seconds to spawn the instance on the hypervisor. [ 1047.137374] env[61985]: DEBUG nova.compute.manager [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1047.137754] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7065e47-b37f-47b1-86dd-674165a04979 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.284805] env[61985]: DEBUG oslo_vmware.api [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Task: {'id': task-936410, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259468} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.284918] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1047.285134] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1047.285892] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1047.286322] env[61985]: INFO nova.compute.manager [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1047.286824] env[61985]: DEBUG oslo.service.loopingcall [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.286998] env[61985]: DEBUG nova.compute.manager [-] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1047.287202] env[61985]: DEBUG nova.network.neutron [-] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1047.419346] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.419619] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.421747] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.422043] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.422245] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.426044] env[61985]: INFO nova.compute.manager [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Terminating instance [ 1047.430821] env[61985]: DEBUG nova.compute.manager [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1047.431245] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1047.432440] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa223928-01ac-4130-9a3e-bbb2fd100e29 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.441899] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1047.442374] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d835d805-453e-402a-8583-7bce85479df9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.450365] env[61985]: DEBUG oslo_vmware.api [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1047.450365] env[61985]: value = "task-936411" [ 1047.450365] env[61985]: _type = "Task" [ 1047.450365] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.460408] env[61985]: DEBUG oslo_vmware.api [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936411, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.568855] env[61985]: DEBUG nova.objects.base [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Object Instance<87006fbf-ea90-4d9a-88af-001de424ac14> lazy-loaded attributes: resources,numa_topology {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1047.658063] env[61985]: INFO nova.compute.manager [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Took 27.42 seconds to build instance. [ 1047.711604] env[61985]: DEBUG nova.network.neutron [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updated VIF entry in instance network info cache for port 88a63ff6-55f6-4b8b-ba9e-188245f7ea87. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1047.712162] env[61985]: DEBUG nova.network.neutron [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updating instance_info_cache with network_info: [{"id": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "address": "fa:16:3e:16:ec:4d", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88a63ff6-55", "ovs_interfaceid": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.878656] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85700ef-9d94-40ca-86ea-474f920c2b18 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.888785] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7601e133-49ec-451d-bb4a-cc1c09e30127 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.924389] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24757c1-daac-4d43-a9dc-e3044b043cd2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.948668] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0575a1-4953-4971-80f4-46cf1dc1a7aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.974569] env[61985]: DEBUG oslo_vmware.api [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936411, 'name': PowerOffVM_Task, 'duration_secs': 0.257828} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.975078] env[61985]: DEBUG nova.compute.provider_tree [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.977525] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1047.977798] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1047.978091] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-426fe9cf-8ec6-49d2-b8e4-d86dff53cd2c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.010031] env[61985]: DEBUG nova.compute.manager [req-687781f2-be21-4e6d-9ce6-a59ae4643f6b req-ac958100-c57a-4ba0-9435-d5fb3ac5349b service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Received event network-vif-deleted-6ea4b2df-e4c4-48a4-9af2-fe4606759004 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1048.010268] env[61985]: INFO nova.compute.manager [req-687781f2-be21-4e6d-9ce6-a59ae4643f6b req-ac958100-c57a-4ba0-9435-d5fb3ac5349b service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Neutron deleted interface 6ea4b2df-e4c4-48a4-9af2-fe4606759004; detaching it from the instance and deleting it from the info cache [ 1048.010900] env[61985]: DEBUG nova.network.neutron [req-687781f2-be21-4e6d-9ce6-a59ae4643f6b req-ac958100-c57a-4ba0-9435-d5fb3ac5349b service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.050931] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1048.051024] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1048.051606] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleting the datastore file [datastore1] 5cd5e044-b0da-4564-8c2c-e894eb29a74c {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1048.051606] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40d1a55d-f107-4624-9b91-61e7d60aef12 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.062765] env[61985]: DEBUG oslo_vmware.api [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for the task: (returnval){ [ 1048.062765] env[61985]: value = "task-936413" [ 1048.062765] env[61985]: _type = "Task" [ 1048.062765] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.076023] env[61985]: DEBUG oslo_vmware.api [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936413, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.162659] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2110fc12-2095-47dc-b957-abc2c987bfa2 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.934s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.177870] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "8eea60eb-93cf-4701-9ba3-0805691ca016" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.178243] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.216749] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b0b09a0-c75d-4996-9179-5837f8bf8d2f req-023a0c1f-9a6a-4f9e-ae93-cb75dfb0090b service nova] Releasing lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.230013] env[61985]: DEBUG nova.network.neutron [-] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.479694] env[61985]: DEBUG nova.scheduler.client.report [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.513919] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c78c5406-ae81-4580-8e58-9d6474e5fdcc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.529523] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a0b025-5ecb-4ed9-b985-515ab6083d15 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.564933] env[61985]: DEBUG nova.compute.manager [req-687781f2-be21-4e6d-9ce6-a59ae4643f6b req-ac958100-c57a-4ba0-9435-d5fb3ac5349b service nova] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Detach interface failed, port_id=6ea4b2df-e4c4-48a4-9af2-fe4606759004, reason: Instance 761d5954-a5ca-4459-a1d6-bfc59b284bf4 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1048.575980] env[61985]: DEBUG oslo_vmware.api [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Task: {'id': task-936413, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.424896} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.576130] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1048.576317] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1048.576738] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1048.576738] env[61985]: INFO nova.compute.manager [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1048.576917] env[61985]: DEBUG oslo.service.loopingcall [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.579242] env[61985]: DEBUG nova.compute.manager [-] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1048.579242] env[61985]: DEBUG nova.network.neutron [-] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1048.624009] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "decfe088-90d6-49f9-9eae-7aa461219eaf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.624285] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "decfe088-90d6-49f9-9eae-7aa461219eaf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.683489] env[61985]: DEBUG nova.compute.manager [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1048.733988] env[61985]: INFO nova.compute.manager [-] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Took 1.44 seconds to deallocate network for instance. [ 1048.990811] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.433s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.994041] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.539s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.131349] env[61985]: DEBUG nova.compute.manager [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1049.208223] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.240948] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.503185] env[61985]: INFO nova.compute.claims [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1049.512820] env[61985]: DEBUG oslo_concurrency.lockutils [None req-2e3e030f-b80c-4375-b9f5-fcb8d7ecfe70 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "87006fbf-ea90-4d9a-88af-001de424ac14" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 28.017s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.514137] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "87006fbf-ea90-4d9a-88af-001de424ac14" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 3.509s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.514426] env[61985]: INFO nova.compute.manager [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Unshelving [ 1049.614750] env[61985]: DEBUG nova.network.neutron [-] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.654035] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.866434] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "9079e605-e865-4eb1-9857-03769e4b5ba6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.866666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.014136] env[61985]: INFO nova.compute.resource_tracker [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating resource usage from migration 3e248959-8537-41af-b11e-32f60bab4249 [ 1050.116997] env[61985]: INFO nova.compute.manager [-] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Took 1.54 seconds to deallocate network for instance. [ 1050.138102] env[61985]: DEBUG nova.compute.manager [req-4899230f-c2ec-4783-9e6c-46b7e33b29b5 req-2727983c-5997-4679-a4d7-de652aa00662 service nova] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Received event network-vif-deleted-6d05ecf8-3c82-464f-be11-a7e2766f9351 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1050.331476] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd532ee5-4f09-4090-9897-d970d527bd2e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.341561] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1b576a-97d7-44a5-907e-d39e1b11b5e4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.380896] env[61985]: DEBUG nova.compute.manager [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1050.384886] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070baaa9-b12a-46d7-b21b-c72fff81eaae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.394121] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf57cd4-89f2-4d8b-a06c-d1234df4e3d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.410227] env[61985]: DEBUG nova.compute.provider_tree [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.541342] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.631440] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.698071] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "44361f7b-4609-476f-b4a9-58a7851e6e92" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.698071] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.698071] env[61985]: DEBUG nova.compute.manager [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.698071] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef2f819-357b-40d4-a415-57291bcee352 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.706058] env[61985]: DEBUG nova.compute.manager [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61985) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3374}} [ 1050.707739] env[61985]: DEBUG nova.objects.instance [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'flavor' on Instance uuid 44361f7b-4609-476f-b4a9-58a7851e6e92 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.909074] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.913284] env[61985]: DEBUG nova.scheduler.client.report [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1051.212912] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1051.213633] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bde04b75-dc03-453a-b0ca-b16f45db63ef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.222418] env[61985]: DEBUG oslo_vmware.api [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1051.222418] env[61985]: value = "task-936414" [ 1051.222418] env[61985]: _type = "Task" [ 1051.222418] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.236643] env[61985]: DEBUG oslo_vmware.api [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936414, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.418222] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.424s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.418456] env[61985]: INFO nova.compute.manager [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Migrating [ 1051.427210] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.874s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.428827] env[61985]: INFO nova.compute.claims [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1051.735662] env[61985]: DEBUG oslo_vmware.api [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936414, 'name': PowerOffVM_Task, 'duration_secs': 0.178803} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.736074] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1051.736331] env[61985]: DEBUG nova.compute.manager [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.737430] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087010f8-eb50-41d6-922a-4eed9e3729ec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.941090] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.943891] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.943891] env[61985]: DEBUG nova.network.neutron [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1052.255468] env[61985]: DEBUG oslo_concurrency.lockutils [None req-73894e07-2083-4041-bc8b-e3b34ab90b49 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.559s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.693090] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279a2fc0-9ed6-488a-a1ee-280f0770806b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.701323] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bdbd35-55ff-4a65-90e3-f941b6d5a415 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.733983] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b92a4316-f0d9-4eea-a3f3-7795adab0bff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.742748] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635f3211-fb32-4f29-bbcf-71ae3a16c750 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.758212] env[61985]: DEBUG nova.compute.provider_tree [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.847354] env[61985]: DEBUG nova.network.neutron [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance_info_cache with network_info: [{"id": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "address": "fa:16:3e:35:b3:a2", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88467e26-c1", "ovs_interfaceid": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.261372] env[61985]: DEBUG nova.scheduler.client.report [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1053.351720] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.685222] env[61985]: DEBUG nova.objects.instance [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'flavor' on Instance uuid 44361f7b-4609-476f-b4a9-58a7851e6e92 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.766634] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.767185] env[61985]: DEBUG nova.compute.manager [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1053.774024] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.972s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.774024] env[61985]: INFO nova.compute.claims [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.192920] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.192920] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.192920] env[61985]: DEBUG nova.network.neutron [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1054.192920] env[61985]: DEBUG nova.objects.instance [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'info_cache' on Instance uuid 44361f7b-4609-476f-b4a9-58a7851e6e92 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.281984] env[61985]: DEBUG nova.compute.utils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1054.289050] env[61985]: DEBUG nova.compute.manager [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1054.289050] env[61985]: DEBUG nova.network.neutron [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1054.408853] env[61985]: DEBUG nova.policy [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faa19f385cd0426d8565e31e286800d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5625013599b44418bd56eb604e14be58', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1054.696203] env[61985]: DEBUG nova.objects.base [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Object Instance<44361f7b-4609-476f-b4a9-58a7851e6e92> lazy-loaded attributes: flavor,info_cache {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1054.788361] env[61985]: DEBUG nova.compute.manager [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1054.874509] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba350f1-7af4-4881-bffc-cd457a08c2e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.905048] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance '8c4a1fe4-4ded-4064-8261-24826181a68c' progress to 0 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1055.019233] env[61985]: DEBUG nova.network.neutron [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Successfully created port: 96ab3107-6606-4dd8-9b69-7d2445a81994 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1055.078187] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585725c8-67a8-4773-ab33-3bbb3a945b0d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.088930] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6371fb9-29cd-4b64-98d7-efee6f73dc4a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.123552] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b3e359-c50b-4c62-8d0d-da2cee144587 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.132596] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a73d848-6883-40c1-8fce-404d7ca689cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.146389] env[61985]: DEBUG nova.compute.provider_tree [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.414414] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1055.414733] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b23539fa-6aec-4344-95d9-155abe08c14d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.423602] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1055.423602] env[61985]: value = "task-936415" [ 1055.423602] env[61985]: _type = "Task" [ 1055.423602] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.432823] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936415, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.433929] env[61985]: DEBUG nova.network.neutron [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance_info_cache with network_info: [{"id": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "address": "fa:16:3e:13:20:58", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35f2970-0b", "ovs_interfaceid": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.650873] env[61985]: DEBUG nova.scheduler.client.report [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.800496] env[61985]: DEBUG nova.compute.manager [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1055.830786] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1055.831058] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1055.831229] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.831461] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1055.831656] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.831857] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1055.832336] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1055.832336] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1055.832446] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1055.832587] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1055.832765] env[61985]: DEBUG nova.virt.hardware [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.833827] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadc5587-f39d-48c2-94f9-244f6c14e66b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.842417] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0041f89-8692-42c4-92cc-57706d775689 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.937173] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.939548] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936415, 'name': PowerOffVM_Task, 'duration_secs': 0.227822} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.939849] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1055.940168] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance '8c4a1fe4-4ded-4064-8261-24826181a68c' progress to 17 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1056.159209] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.159697] env[61985]: DEBUG nova.compute.manager [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1056.162600] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.955s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.164312] env[61985]: INFO nova.compute.claims [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1056.306127] env[61985]: DEBUG oslo_concurrency.lockutils [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "interface-8db82979-e527-4c63-8ce8-fe1456f7708c-0822504d-b73d-40b5-a7a2-4770f9d2f190" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.306127] env[61985]: DEBUG oslo_concurrency.lockutils [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-8db82979-e527-4c63-8ce8-fe1456f7708c-0822504d-b73d-40b5-a7a2-4770f9d2f190" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.306127] env[61985]: DEBUG nova.objects.instance [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'flavor' on Instance uuid 8db82979-e527-4c63-8ce8-fe1456f7708c {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.441981] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1056.441981] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b071a0c-547b-4b4c-be00-c470da2ab755 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.445350] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1056.445678] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1056.445862] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1056.446069] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1056.446227] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1056.446394] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1056.446580] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1056.446737] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1056.447561] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1056.447561] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1056.447561] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.452497] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c4f75ae-168e-46a4-a0af-0c8f48e3ceed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.468559] env[61985]: DEBUG oslo_vmware.api [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1056.468559] env[61985]: value = "task-936416" [ 1056.468559] env[61985]: _type = "Task" [ 1056.468559] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.469897] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1056.469897] env[61985]: value = "task-936417" [ 1056.469897] env[61985]: _type = "Task" [ 1056.469897] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.481453] env[61985]: DEBUG oslo_vmware.api [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.484278] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936417, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.669293] env[61985]: DEBUG nova.compute.utils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1056.673062] env[61985]: DEBUG nova.compute.manager [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1056.673268] env[61985]: DEBUG nova.network.neutron [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1056.721242] env[61985]: DEBUG nova.policy [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b5abdace901430cb1e431c9a933161e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bdb5f76b9b24a7ba4800a88482847da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1056.975090] env[61985]: DEBUG nova.objects.instance [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'pci_requests' on Instance uuid 8db82979-e527-4c63-8ce8-fe1456f7708c {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.992500] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936417, 'name': ReconfigVM_Task, 'duration_secs': 0.184432} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.992756] env[61985]: DEBUG oslo_vmware.api [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936416, 'name': PowerOnVM_Task, 'duration_secs': 0.40267} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.993302] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance '8c4a1fe4-4ded-4064-8261-24826181a68c' progress to 33 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1056.997516] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1056.997746] env[61985]: DEBUG nova.compute.manager [None req-f78784e1-884b-4dd7-b927-3c4a698622a7 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.999103] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf71d14-7bbd-45c0-8e4c-e9f5112b6e9d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.017229] env[61985]: DEBUG nova.network.neutron [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Successfully updated port: 96ab3107-6606-4dd8-9b69-7d2445a81994 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.019487] env[61985]: DEBUG nova.compute.manager [req-29364452-7f1f-4d74-ab52-2d931cf8c19b req-ee950e71-79ac-449e-9330-8353055b7981 service nova] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Received event network-vif-plugged-96ab3107-6606-4dd8-9b69-7d2445a81994 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1057.019745] env[61985]: DEBUG oslo_concurrency.lockutils [req-29364452-7f1f-4d74-ab52-2d931cf8c19b req-ee950e71-79ac-449e-9330-8353055b7981 service nova] Acquiring lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.019942] env[61985]: DEBUG oslo_concurrency.lockutils [req-29364452-7f1f-4d74-ab52-2d931cf8c19b req-ee950e71-79ac-449e-9330-8353055b7981 service nova] Lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.020152] env[61985]: DEBUG oslo_concurrency.lockutils [req-29364452-7f1f-4d74-ab52-2d931cf8c19b req-ee950e71-79ac-449e-9330-8353055b7981 service nova] Lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.020328] env[61985]: DEBUG nova.compute.manager [req-29364452-7f1f-4d74-ab52-2d931cf8c19b req-ee950e71-79ac-449e-9330-8353055b7981 service nova] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] No waiting events found dispatching network-vif-plugged-96ab3107-6606-4dd8-9b69-7d2445a81994 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1057.020497] env[61985]: WARNING nova.compute.manager [req-29364452-7f1f-4d74-ab52-2d931cf8c19b req-ee950e71-79ac-449e-9330-8353055b7981 service nova] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Received unexpected event network-vif-plugged-96ab3107-6606-4dd8-9b69-7d2445a81994 for instance with vm_state building and task_state spawning. [ 1057.119642] env[61985]: DEBUG nova.network.neutron [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Successfully created port: 52d657fa-a084-485b-858a-368bb739f4fc {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1057.173580] env[61985]: DEBUG nova.compute.manager [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1057.390464] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06409fb-acea-4412-9917-7602084f241e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.398918] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a811172-cf12-4ee9-aee3-772038480482 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.429030] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b625b1-c3d8-4526-8593-8f1ac83c2f61 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.436625] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a50d43-b891-464b-b9ea-a78a9594c0e8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.450543] env[61985]: DEBUG nova.compute.provider_tree [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.481522] env[61985]: DEBUG nova.objects.base [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Object Instance<8db82979-e527-4c63-8ce8-fe1456f7708c> lazy-loaded attributes: flavor,pci_requests {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1057.481756] env[61985]: DEBUG nova.network.neutron [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1057.500570] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1057.500825] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1057.500990] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.501198] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1057.501355] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.501509] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1057.501716] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1057.501892] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1057.502067] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1057.502242] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1057.502419] env[61985]: DEBUG nova.virt.hardware [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1057.508050] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Reconfiguring VM instance instance-0000005c to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1057.510381] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fbd83bd-8a0a-4269-997d-42139526649e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.528834] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "refresh_cache-65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.528911] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "refresh_cache-65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.529126] env[61985]: DEBUG nova.network.neutron [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1057.538786] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1057.538786] env[61985]: value = "task-936418" [ 1057.538786] env[61985]: _type = "Task" [ 1057.538786] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.549585] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936418, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.596863] env[61985]: DEBUG nova.policy [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a78dbc667d24c819e6447740fc2d8e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2dbceebf07c43ac90ed80f8a09294b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1057.954183] env[61985]: DEBUG nova.scheduler.client.report [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.050544] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936418, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.061406] env[61985]: DEBUG nova.network.neutron [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1058.186226] env[61985]: DEBUG nova.compute.manager [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1058.203887] env[61985]: DEBUG nova.network.neutron [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Updating instance_info_cache with network_info: [{"id": "96ab3107-6606-4dd8-9b69-7d2445a81994", "address": "fa:16:3e:82:c2:ea", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96ab3107-66", "ovs_interfaceid": "96ab3107-6606-4dd8-9b69-7d2445a81994", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.216980] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1058.218030] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1058.218030] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.218030] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1058.218030] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.218030] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1058.218233] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1058.218409] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1058.218609] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1058.218919] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1058.219131] env[61985]: DEBUG nova.virt.hardware [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.220097] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f27099-0296-47a2-9762-83c2805ebbcf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.230725] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa627a8-dfe7-4c1f-8b3a-ac64b08114e6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.459415] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.460367] env[61985]: DEBUG nova.compute.manager [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1058.462721] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.222s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.462962] env[61985]: DEBUG nova.objects.instance [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lazy-loading 'resources' on Instance uuid 761d5954-a5ca-4459-a1d6-bfc59b284bf4 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.552052] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936418, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.624751] env[61985]: DEBUG nova.network.neutron [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Successfully updated port: 52d657fa-a084-485b-858a-368bb739f4fc {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1058.710208] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "refresh_cache-65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.710208] env[61985]: DEBUG nova.compute.manager [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Instance network_info: |[{"id": "96ab3107-6606-4dd8-9b69-7d2445a81994", "address": "fa:16:3e:82:c2:ea", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96ab3107-66", "ovs_interfaceid": "96ab3107-6606-4dd8-9b69-7d2445a81994", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1058.710648] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:c2:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7908211b-df93-467b-87a8-3c3d29b03de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96ab3107-6606-4dd8-9b69-7d2445a81994', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.721607] env[61985]: DEBUG oslo.service.loopingcall [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.721607] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1058.721607] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35c63d1a-2e3b-48a4-ad86-2579de6c95b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.745169] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.745169] env[61985]: value = "task-936419" [ 1058.745169] env[61985]: _type = "Task" [ 1058.745169] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.754097] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936419, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.776030] env[61985]: DEBUG nova.compute.manager [req-203db1ee-9fa7-4152-b5e5-3fe03b5e049e req-d3ecb8d6-22e6-4e6a-afa2-78172148461a service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Received event network-vif-plugged-52d657fa-a084-485b-858a-368bb739f4fc {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1058.776863] env[61985]: DEBUG oslo_concurrency.lockutils [req-203db1ee-9fa7-4152-b5e5-3fe03b5e049e req-d3ecb8d6-22e6-4e6a-afa2-78172148461a service nova] Acquiring lock "036fae78-2af0-49f1-85fa-245c29f35ebc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.776863] env[61985]: DEBUG oslo_concurrency.lockutils [req-203db1ee-9fa7-4152-b5e5-3fe03b5e049e req-d3ecb8d6-22e6-4e6a-afa2-78172148461a service nova] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.776863] env[61985]: DEBUG oslo_concurrency.lockutils [req-203db1ee-9fa7-4152-b5e5-3fe03b5e049e req-d3ecb8d6-22e6-4e6a-afa2-78172148461a service nova] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.776863] env[61985]: DEBUG nova.compute.manager [req-203db1ee-9fa7-4152-b5e5-3fe03b5e049e req-d3ecb8d6-22e6-4e6a-afa2-78172148461a service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] No waiting events found dispatching network-vif-plugged-52d657fa-a084-485b-858a-368bb739f4fc {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1058.777015] env[61985]: WARNING nova.compute.manager [req-203db1ee-9fa7-4152-b5e5-3fe03b5e049e req-d3ecb8d6-22e6-4e6a-afa2-78172148461a service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Received unexpected event network-vif-plugged-52d657fa-a084-485b-858a-368bb739f4fc for instance with vm_state building and task_state spawning. [ 1058.965774] env[61985]: DEBUG nova.compute.utils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1058.967277] env[61985]: DEBUG nova.compute.manager [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1058.967458] env[61985]: DEBUG nova.network.neutron [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1059.025522] env[61985]: DEBUG nova.policy [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc86b7a24ed94efbae2dc437e4fef389', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '60ecaef69401418eac6c655c1b7df5be', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1059.055999] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936418, 'name': ReconfigVM_Task, 'duration_secs': 1.188921} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.059216] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Reconfigured VM instance instance-0000005c to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1059.060419] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a2ff23-b376-4d8f-8d05-2fbafc5529ba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.084182] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 8c4a1fe4-4ded-4064-8261-24826181a68c/8c4a1fe4-4ded-4064-8261-24826181a68c.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.087208] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67d0f32e-f76c-4e25-9c64-7ed1d89d5ca3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.106838] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1059.106838] env[61985]: value = "task-936420" [ 1059.106838] env[61985]: _type = "Task" [ 1059.106838] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.118698] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936420, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.128795] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "refresh_cache-036fae78-2af0-49f1-85fa-245c29f35ebc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.128795] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "refresh_cache-036fae78-2af0-49f1-85fa-245c29f35ebc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.129386] env[61985]: DEBUG nova.network.neutron [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1059.233272] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd2b57b-157c-4c86-b46f-71aec40c5a1c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.242248] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d6bed4-4c92-401b-9acc-b185c83e76e6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.275022] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34474600-afa3-4985-a05e-009d42432208 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.279473] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936419, 'name': CreateVM_Task, 'duration_secs': 0.369681} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.280039] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1059.280791] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.280969] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.281328] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1059.281593] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62905e9a-4342-4a6b-98a6-9ed5b2b275ce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.288589] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d93b84-3254-4dd7-b1c2-674828870a24 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.293926] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1059.293926] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5203e542-553e-a2ba-6139-cb03e0bd3161" [ 1059.293926] env[61985]: _type = "Task" [ 1059.293926] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.305271] env[61985]: DEBUG nova.compute.provider_tree [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.312580] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5203e542-553e-a2ba-6139-cb03e0bd3161, 'name': SearchDatastore_Task, 'duration_secs': 0.011103} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.312897] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.313167] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1059.313416] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.313578] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.313772] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1059.314064] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ac37117-cf6a-4cec-a68f-4a9274e88a8f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.325104] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1059.325104] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1059.325988] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5a28799-6568-41e4-8fa2-9ddfff4adafe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.333440] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1059.333440] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cb81e1-a2ae-1ba5-3cef-af7c9d89b67c" [ 1059.333440] env[61985]: _type = "Task" [ 1059.333440] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.343843] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cb81e1-a2ae-1ba5-3cef-af7c9d89b67c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.476567] env[61985]: DEBUG nova.compute.manager [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1059.495791] env[61985]: DEBUG nova.network.neutron [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Successfully created port: 01e88a1a-79a9-4c6f-a939-607ae4ecd7f2 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1059.512671] env[61985]: DEBUG nova.compute.manager [req-6db875a5-061c-49d0-b23f-114133e8191b req-7fdc11c5-cbc8-46c8-971e-9d219d519f52 service nova] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Received event network-changed-96ab3107-6606-4dd8-9b69-7d2445a81994 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1059.512871] env[61985]: DEBUG nova.compute.manager [req-6db875a5-061c-49d0-b23f-114133e8191b req-7fdc11c5-cbc8-46c8-971e-9d219d519f52 service nova] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Refreshing instance network info cache due to event network-changed-96ab3107-6606-4dd8-9b69-7d2445a81994. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1059.513103] env[61985]: DEBUG oslo_concurrency.lockutils [req-6db875a5-061c-49d0-b23f-114133e8191b req-7fdc11c5-cbc8-46c8-971e-9d219d519f52 service nova] Acquiring lock "refresh_cache-65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.513254] env[61985]: DEBUG oslo_concurrency.lockutils [req-6db875a5-061c-49d0-b23f-114133e8191b req-7fdc11c5-cbc8-46c8-971e-9d219d519f52 service nova] Acquired lock "refresh_cache-65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.513488] env[61985]: DEBUG nova.network.neutron [req-6db875a5-061c-49d0-b23f-114133e8191b req-7fdc11c5-cbc8-46c8-971e-9d219d519f52 service nova] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Refreshing network info cache for port 96ab3107-6606-4dd8-9b69-7d2445a81994 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1059.617378] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936420, 'name': ReconfigVM_Task, 'duration_secs': 0.278723} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.617710] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 8c4a1fe4-4ded-4064-8261-24826181a68c/8c4a1fe4-4ded-4064-8261-24826181a68c.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1059.617999] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance '8c4a1fe4-4ded-4064-8261-24826181a68c' progress to 50 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1059.680396] env[61985]: DEBUG nova.network.neutron [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1059.751726] env[61985]: DEBUG nova.network.neutron [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Successfully updated port: 0822504d-b73d-40b5-a7a2-4770f9d2f190 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1059.808385] env[61985]: DEBUG nova.scheduler.client.report [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.844480] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52cb81e1-a2ae-1ba5-3cef-af7c9d89b67c, 'name': SearchDatastore_Task, 'duration_secs': 0.029473} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.845748] env[61985]: DEBUG nova.network.neutron [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Successfully created port: c4f26df9-9705-48e3-8414-595fed68fdee {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1059.847422] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0c5e5b6-ae0e-4e40-b37e-f6afdc7cd2ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.854348] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1059.854348] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524d5f98-76b6-96ab-9224-2724ccd25614" [ 1059.854348] env[61985]: _type = "Task" [ 1059.854348] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.863527] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524d5f98-76b6-96ab-9224-2724ccd25614, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.884422] env[61985]: DEBUG nova.network.neutron [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Updating instance_info_cache with network_info: [{"id": "52d657fa-a084-485b-858a-368bb739f4fc", "address": "fa:16:3e:8f:ab:a9", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52d657fa-a0", "ovs_interfaceid": "52d657fa-a084-485b-858a-368bb739f4fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.127169] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3369a9b3-b324-419a-bfc5-92c2445b47e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.153931] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f553d90-d102-40b1-a6cd-64682b111fe3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.173315] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance '8c4a1fe4-4ded-4064-8261-24826181a68c' progress to 67 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1060.219753] env[61985]: DEBUG nova.network.neutron [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Successfully created port: 6a2e1f2a-6455-4354-83df-19ac0b8f17db {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1060.254358] env[61985]: DEBUG oslo_concurrency.lockutils [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.254561] env[61985]: DEBUG oslo_concurrency.lockutils [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.254748] env[61985]: DEBUG nova.network.neutron [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1060.313786] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.315844] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.662s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.317476] env[61985]: INFO nova.compute.claims [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1060.335028] env[61985]: INFO nova.scheduler.client.report [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Deleted allocations for instance 761d5954-a5ca-4459-a1d6-bfc59b284bf4 [ 1060.364635] env[61985]: DEBUG nova.network.neutron [req-6db875a5-061c-49d0-b23f-114133e8191b req-7fdc11c5-cbc8-46c8-971e-9d219d519f52 service nova] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Updated VIF entry in instance network info cache for port 96ab3107-6606-4dd8-9b69-7d2445a81994. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1060.364994] env[61985]: DEBUG nova.network.neutron [req-6db875a5-061c-49d0-b23f-114133e8191b req-7fdc11c5-cbc8-46c8-971e-9d219d519f52 service nova] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Updating instance_info_cache with network_info: [{"id": "96ab3107-6606-4dd8-9b69-7d2445a81994", "address": "fa:16:3e:82:c2:ea", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96ab3107-66", "ovs_interfaceid": "96ab3107-6606-4dd8-9b69-7d2445a81994", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.376564] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524d5f98-76b6-96ab-9224-2724ccd25614, 'name': SearchDatastore_Task, 'duration_secs': 0.008872} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.376843] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.377132] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507/65caa6d1-6e6e-4ab3-bb80-4c7e3b380507.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1060.377645] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-97e36292-e14b-4255-adc7-6a4004cf7885 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.386371] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1060.386371] env[61985]: value = "task-936421" [ 1060.386371] env[61985]: _type = "Task" [ 1060.386371] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.389832] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "refresh_cache-036fae78-2af0-49f1-85fa-245c29f35ebc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.390166] env[61985]: DEBUG nova.compute.manager [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Instance network_info: |[{"id": "52d657fa-a084-485b-858a-368bb739f4fc", "address": "fa:16:3e:8f:ab:a9", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52d657fa-a0", "ovs_interfaceid": "52d657fa-a084-485b-858a-368bb739f4fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1060.390924] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:ab:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52d657fa-a084-485b-858a-368bb739f4fc', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1060.398732] env[61985]: DEBUG oslo.service.loopingcall [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.399327] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1060.399950] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f2451de-cfd2-4738-97f5-8cec7ebca365 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.419245] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936421, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.425716] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1060.425716] env[61985]: value = "task-936422" [ 1060.425716] env[61985]: _type = "Task" [ 1060.425716] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.435404] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936422, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.488473] env[61985]: DEBUG nova.compute.manager [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1060.517015] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1060.517015] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1060.517015] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1060.517015] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1060.517015] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1060.517015] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1060.517015] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1060.517015] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1060.517015] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1060.517397] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1060.517397] env[61985]: DEBUG nova.virt.hardware [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.518569] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9025b968-6330-41b1-9ce0-91cbfffff23c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.528346] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56d3ff2-807c-4515-9a19-fa91bcfac9b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.719108] env[61985]: DEBUG nova.network.neutron [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Port 88467e26-c1c7-4b6a-ab2b-9d86b552c0c1 binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1060.802701] env[61985]: WARNING nova.network.neutron [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] 5014dbd4-285f-43e0-ad95-2107c34e808a already exists in list: networks containing: ['5014dbd4-285f-43e0-ad95-2107c34e808a']. ignoring it [ 1060.827225] env[61985]: DEBUG nova.compute.manager [req-a5ae4079-0f47-4a48-a461-294329e39098 req-d8841822-6fde-4ab1-90e9-3df46c02ac56 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Received event network-changed-52d657fa-a084-485b-858a-368bb739f4fc {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1060.827430] env[61985]: DEBUG nova.compute.manager [req-a5ae4079-0f47-4a48-a461-294329e39098 req-d8841822-6fde-4ab1-90e9-3df46c02ac56 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Refreshing instance network info cache due to event network-changed-52d657fa-a084-485b-858a-368bb739f4fc. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1060.827723] env[61985]: DEBUG oslo_concurrency.lockutils [req-a5ae4079-0f47-4a48-a461-294329e39098 req-d8841822-6fde-4ab1-90e9-3df46c02ac56 service nova] Acquiring lock "refresh_cache-036fae78-2af0-49f1-85fa-245c29f35ebc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.827892] env[61985]: DEBUG oslo_concurrency.lockutils [req-a5ae4079-0f47-4a48-a461-294329e39098 req-d8841822-6fde-4ab1-90e9-3df46c02ac56 service nova] Acquired lock "refresh_cache-036fae78-2af0-49f1-85fa-245c29f35ebc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.828476] env[61985]: DEBUG nova.network.neutron [req-a5ae4079-0f47-4a48-a461-294329e39098 req-d8841822-6fde-4ab1-90e9-3df46c02ac56 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Refreshing network info cache for port 52d657fa-a084-485b-858a-368bb739f4fc {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1060.849907] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e9a47ebd-5a98-44bc-852a-754c034294e6 tempest-ServerRescueNegativeTestJSON-1357938369 tempest-ServerRescueNegativeTestJSON-1357938369-project-member] Lock "761d5954-a5ca-4459-a1d6-bfc59b284bf4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.721s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.869159] env[61985]: DEBUG oslo_concurrency.lockutils [req-6db875a5-061c-49d0-b23f-114133e8191b req-7fdc11c5-cbc8-46c8-971e-9d219d519f52 service nova] Releasing lock "refresh_cache-65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.896361] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936421, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444306} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.899435] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507/65caa6d1-6e6e-4ab3-bb80-4c7e3b380507.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1060.899888] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1060.900738] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e0f2dbac-64f0-4a23-896e-46edec3f9d09 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.908155] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1060.908155] env[61985]: value = "task-936423" [ 1060.908155] env[61985]: _type = "Task" [ 1060.908155] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.915881] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936423, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.934803] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936422, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.090764] env[61985]: DEBUG nova.network.neutron [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updating instance_info_cache with network_info: [{"id": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "address": "fa:16:3e:16:ec:4d", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88a63ff6-55", "ovs_interfaceid": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0822504d-b73d-40b5-a7a2-4770f9d2f190", "address": "fa:16:3e:ab:c2:ce", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0822504d-b7", "ovs_interfaceid": "0822504d-b73d-40b5-a7a2-4770f9d2f190", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.418390] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936423, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.437591] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936422, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.549434] env[61985]: DEBUG nova.compute.manager [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Received event network-vif-plugged-0822504d-b73d-40b5-a7a2-4770f9d2f190 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1061.549434] env[61985]: DEBUG oslo_concurrency.lockutils [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] Acquiring lock "8db82979-e527-4c63-8ce8-fe1456f7708c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.549434] env[61985]: DEBUG oslo_concurrency.lockutils [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] Lock "8db82979-e527-4c63-8ce8-fe1456f7708c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.549748] env[61985]: DEBUG oslo_concurrency.lockutils [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] Lock "8db82979-e527-4c63-8ce8-fe1456f7708c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.549748] env[61985]: DEBUG nova.compute.manager [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] No waiting events found dispatching network-vif-plugged-0822504d-b73d-40b5-a7a2-4770f9d2f190 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1061.550827] env[61985]: WARNING nova.compute.manager [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Received unexpected event network-vif-plugged-0822504d-b73d-40b5-a7a2-4770f9d2f190 for instance with vm_state active and task_state None. [ 1061.550827] env[61985]: DEBUG nova.compute.manager [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Received event network-changed-0822504d-b73d-40b5-a7a2-4770f9d2f190 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1061.550827] env[61985]: DEBUG nova.compute.manager [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Refreshing instance network info cache due to event network-changed-0822504d-b73d-40b5-a7a2-4770f9d2f190. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1061.550827] env[61985]: DEBUG oslo_concurrency.lockutils [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] Acquiring lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.566186] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0faa2f-0d0a-44f8-9fa8-dbd78b760093 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.574331] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9f73fa-4b76-4c1b-ab51-38ad460c4629 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.578074] env[61985]: DEBUG nova.network.neutron [req-a5ae4079-0f47-4a48-a461-294329e39098 req-d8841822-6fde-4ab1-90e9-3df46c02ac56 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Updated VIF entry in instance network info cache for port 52d657fa-a084-485b-858a-368bb739f4fc. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1061.578410] env[61985]: DEBUG nova.network.neutron [req-a5ae4079-0f47-4a48-a461-294329e39098 req-d8841822-6fde-4ab1-90e9-3df46c02ac56 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Updating instance_info_cache with network_info: [{"id": "52d657fa-a084-485b-858a-368bb739f4fc", "address": "fa:16:3e:8f:ab:a9", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52d657fa-a0", "ovs_interfaceid": "52d657fa-a084-485b-858a-368bb739f4fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.609052] env[61985]: DEBUG oslo_concurrency.lockutils [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.609656] env[61985]: DEBUG oslo_concurrency.lockutils [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.609937] env[61985]: DEBUG oslo_concurrency.lockutils [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.610802] env[61985]: DEBUG oslo_concurrency.lockutils [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] Acquired lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.610996] env[61985]: DEBUG nova.network.neutron [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Refreshing network info cache for port 0822504d-b73d-40b5-a7a2-4770f9d2f190 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1061.612581] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa17154-03f2-4138-8a25-677f282d02b7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.618022] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c01b245-5948-43e7-9b67-93b7d59b505b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.637586] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.637861] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.638044] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.638238] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.638391] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.638544] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.638790] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.638956] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.639143] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.639314] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.639491] env[61985]: DEBUG nova.virt.hardware [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.646210] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Reconfiguring VM to attach interface {{(pid=61985) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1061.647238] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b553fe20-e041-4446-a915-635f295319b6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.660658] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43759ee-466c-4fd7-881d-bd1352b93378 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.676634] env[61985]: DEBUG nova.compute.provider_tree [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.678952] env[61985]: DEBUG oslo_vmware.api [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1061.678952] env[61985]: value = "task-936424" [ 1061.678952] env[61985]: _type = "Task" [ 1061.678952] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.688529] env[61985]: DEBUG oslo_vmware.api [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936424, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.744454] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "8c4a1fe4-4ded-4064-8261-24826181a68c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.744454] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.744454] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.807550] env[61985]: DEBUG nova.network.neutron [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Successfully updated port: 01e88a1a-79a9-4c6f-a939-607ae4ecd7f2 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1061.918715] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936423, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.877565} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.919066] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.919877] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9f434f-c1a1-4cb3-98f0-917e64e6f7c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.943965] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507/65caa6d1-6e6e-4ab3-bb80-4c7e3b380507.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.947386] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f13be51e-5ca3-4494-bdac-b48483056da8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.967459] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936422, 'name': CreateVM_Task, 'duration_secs': 1.370935} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.968703] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1061.969089] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1061.969089] env[61985]: value = "task-936425" [ 1061.969089] env[61985]: _type = "Task" [ 1061.969089] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.969763] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.969960] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.970293] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1061.970588] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6856ac4-b4db-4786-a8fa-233ddd38d4fe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.981086] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1061.981086] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a909a0-ca21-7f5e-b99a-0ec3ed376386" [ 1061.981086] env[61985]: _type = "Task" [ 1061.981086] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.981342] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936425, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.993639] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a909a0-ca21-7f5e-b99a-0ec3ed376386, 'name': SearchDatastore_Task, 'duration_secs': 0.009522} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.993639] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.993639] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.993639] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.993639] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.993639] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.993639] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d7fa93b1-843b-41dc-af6f-82b6d915c7ef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.000342] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1062.000614] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1062.001260] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaabbad1-587d-4d43-9b50-b222609b01a5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.006601] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1062.006601] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fdeebd-d53a-b5fe-4795-473451567a13" [ 1062.006601] env[61985]: _type = "Task" [ 1062.006601] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.014378] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fdeebd-d53a-b5fe-4795-473451567a13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.081372] env[61985]: DEBUG oslo_concurrency.lockutils [req-a5ae4079-0f47-4a48-a461-294329e39098 req-d8841822-6fde-4ab1-90e9-3df46c02ac56 service nova] Releasing lock "refresh_cache-036fae78-2af0-49f1-85fa-245c29f35ebc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.180993] env[61985]: DEBUG nova.scheduler.client.report [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1062.193761] env[61985]: DEBUG oslo_vmware.api [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936424, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.455967] env[61985]: DEBUG nova.network.neutron [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updated VIF entry in instance network info cache for port 0822504d-b73d-40b5-a7a2-4770f9d2f190. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1062.455967] env[61985]: DEBUG nova.network.neutron [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updating instance_info_cache with network_info: [{"id": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "address": "fa:16:3e:16:ec:4d", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88a63ff6-55", "ovs_interfaceid": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0822504d-b73d-40b5-a7a2-4770f9d2f190", "address": "fa:16:3e:ab:c2:ce", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0822504d-b7", "ovs_interfaceid": "0822504d-b73d-40b5-a7a2-4770f9d2f190", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.483575] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936425, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.520277] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fdeebd-d53a-b5fe-4795-473451567a13, 'name': SearchDatastore_Task, 'duration_secs': 0.010191} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.525111] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10345f34-ebd9-409c-b64b-7b296aa3eec4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.528823] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1062.528823] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52eef113-bfc3-aa6d-017c-3c61e97902dc" [ 1062.528823] env[61985]: _type = "Task" [ 1062.528823] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.538270] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52eef113-bfc3-aa6d-017c-3c61e97902dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.691051] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.691051] env[61985]: DEBUG nova.compute.manager [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1062.692251] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.151s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.692415] env[61985]: DEBUG nova.objects.instance [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lazy-loading 'pci_requests' on Instance uuid 87006fbf-ea90-4d9a-88af-001de424ac14 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.701892] env[61985]: DEBUG oslo_vmware.api [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936424, 'name': ReconfigVM_Task, 'duration_secs': 0.783631} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.701892] env[61985]: DEBUG oslo_concurrency.lockutils [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.701892] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Reconfigured VM to attach interface {{(pid=61985) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1062.796704] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.796912] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.797114] env[61985]: DEBUG nova.network.neutron [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1062.961738] env[61985]: DEBUG oslo_concurrency.lockutils [req-6bdf0453-da01-4126-9c6d-caceef688faa req-c69102a8-e650-4e83-99c1-67dfc6747f38 service nova] Releasing lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.980795] env[61985]: DEBUG nova.compute.manager [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received event network-vif-plugged-01e88a1a-79a9-4c6f-a939-607ae4ecd7f2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1062.981200] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] Acquiring lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.981500] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.981767] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.982210] env[61985]: DEBUG nova.compute.manager [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] No waiting events found dispatching network-vif-plugged-01e88a1a-79a9-4c6f-a939-607ae4ecd7f2 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1062.982535] env[61985]: WARNING nova.compute.manager [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received unexpected event network-vif-plugged-01e88a1a-79a9-4c6f-a939-607ae4ecd7f2 for instance with vm_state building and task_state spawning. [ 1062.982743] env[61985]: DEBUG nova.compute.manager [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received event network-changed-01e88a1a-79a9-4c6f-a939-607ae4ecd7f2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1062.982929] env[61985]: DEBUG nova.compute.manager [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Refreshing instance network info cache due to event network-changed-01e88a1a-79a9-4c6f-a939-607ae4ecd7f2. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1062.983133] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] Acquiring lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.983279] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] Acquired lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.983439] env[61985]: DEBUG nova.network.neutron [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Refreshing network info cache for port 01e88a1a-79a9-4c6f-a939-607ae4ecd7f2 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1062.988832] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936425, 'name': ReconfigVM_Task, 'duration_secs': 0.521746} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.990031] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507/65caa6d1-6e6e-4ab3-bb80-4c7e3b380507.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.991399] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76be1f3d-33e0-49b5-9fff-22c7bc88811c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.998470] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1062.998470] env[61985]: value = "task-936426" [ 1062.998470] env[61985]: _type = "Task" [ 1062.998470] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.007110] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936426, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.039836] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52eef113-bfc3-aa6d-017c-3c61e97902dc, 'name': SearchDatastore_Task, 'duration_secs': 0.015377} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.040155] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.040483] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 036fae78-2af0-49f1-85fa-245c29f35ebc/036fae78-2af0-49f1-85fa-245c29f35ebc.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1063.041011] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1414fc60-8481-494f-9b22-5d93524f0b34 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.047544] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1063.047544] env[61985]: value = "task-936427" [ 1063.047544] env[61985]: _type = "Task" [ 1063.047544] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.056707] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936427, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.198169] env[61985]: DEBUG nova.objects.instance [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lazy-loading 'numa_topology' on Instance uuid 87006fbf-ea90-4d9a-88af-001de424ac14 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.200197] env[61985]: DEBUG nova.compute.utils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1063.201411] env[61985]: DEBUG nova.compute.manager [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1063.201646] env[61985]: DEBUG nova.network.neutron [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1063.207261] env[61985]: DEBUG oslo_concurrency.lockutils [None req-29d9a42b-b392-4bab-a63f-3e80080ec531 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-8db82979-e527-4c63-8ce8-fe1456f7708c-0822504d-b73d-40b5-a7a2-4770f9d2f190" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.903s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.270431] env[61985]: DEBUG nova.policy [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e23fdb33638143409613d9d4f26f199c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe531ec9f60548adb457351e19f8d51b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1063.515836] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936426, 'name': Rename_Task, 'duration_secs': 0.125512} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.515836] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1063.515836] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34ee6941-0f4c-48e9-a6f7-5622a3c21d1e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.522903] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1063.522903] env[61985]: value = "task-936428" [ 1063.522903] env[61985]: _type = "Task" [ 1063.522903] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.532753] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936428, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.558907] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936427, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.568240] env[61985]: DEBUG nova.network.neutron [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1063.707914] env[61985]: DEBUG nova.compute.manager [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1063.716032] env[61985]: INFO nova.compute.claims [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1063.726454] env[61985]: DEBUG nova.network.neutron [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance_info_cache with network_info: [{"id": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "address": "fa:16:3e:35:b3:a2", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88467e26-c1", "ovs_interfaceid": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.747702] env[61985]: DEBUG nova.network.neutron [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Successfully created port: 5a9c3e06-ff99-444a-a9ed-8f508e1854fb {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1063.753325] env[61985]: DEBUG nova.network.neutron [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.044326] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936428, 'name': PowerOnVM_Task} progress is 80%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.059842] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936427, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.804677} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.060129] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 036fae78-2af0-49f1-85fa-245c29f35ebc/036fae78-2af0-49f1-85fa-245c29f35ebc.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1064.060362] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1064.060599] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f673f4a-204c-4583-8a61-3b8541dfe225 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.070418] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1064.070418] env[61985]: value = "task-936429" [ 1064.070418] env[61985]: _type = "Task" [ 1064.070418] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.079887] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936429, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.233197] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.258423] env[61985]: DEBUG oslo_concurrency.lockutils [req-3b36d529-ff82-44de-a5dc-fc35eae93b66 req-1f18d4a5-cc9f-413c-9dc7-c5d4cbe0a5ce service nova] Releasing lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.259615] env[61985]: DEBUG nova.network.neutron [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Successfully updated port: c4f26df9-9705-48e3-8414-595fed68fdee {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.534206] env[61985]: DEBUG oslo_vmware.api [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936428, 'name': PowerOnVM_Task, 'duration_secs': 0.948239} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.534479] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1064.534688] env[61985]: INFO nova.compute.manager [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Took 8.73 seconds to spawn the instance on the hypervisor. [ 1064.534878] env[61985]: DEBUG nova.compute.manager [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1064.535665] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f23d2187-49cd-4949-a5db-ae45f248c8bf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.581573] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936429, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07143} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.581917] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1064.582870] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbfb2d13-4732-4bb6-81c3-8f6dd19ce4c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.605827] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 036fae78-2af0-49f1-85fa-245c29f35ebc/036fae78-2af0-49f1-85fa-245c29f35ebc.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.606162] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-136d55e6-fbd0-4856-8ee5-62775585a072 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.625688] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1064.625688] env[61985]: value = "task-936430" [ 1064.625688] env[61985]: _type = "Task" [ 1064.625688] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.634018] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936430, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.663519] env[61985]: DEBUG oslo_concurrency.lockutils [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "interface-8db82979-e527-4c63-8ce8-fe1456f7708c-0822504d-b73d-40b5-a7a2-4770f9d2f190" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.663652] env[61985]: DEBUG oslo_concurrency.lockutils [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-8db82979-e527-4c63-8ce8-fe1456f7708c-0822504d-b73d-40b5-a7a2-4770f9d2f190" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.725545] env[61985]: DEBUG nova.compute.manager [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1064.759635] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1064.759936] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1064.760129] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.760328] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1064.760482] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.760637] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1064.760986] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1064.761261] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1064.761590] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1064.762567] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1064.762965] env[61985]: DEBUG nova.virt.hardware [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.764391] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea87813b-3b12-4461-8b02-565a4f79c5d7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.776337] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ab09a0-f357-4751-97da-5894d6b82291 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.806303] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f718e5b0-6ab0-4cb6-8846-e7b36d135a69 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.815023] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25c03ac-e4ba-4580-8c25-3c6dd4444333 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.823754] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance '8c4a1fe4-4ded-4064-8261-24826181a68c' progress to 83 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1064.990566] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f6dc513-c7da-433d-9df9-a7fcbc07e6b7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.998636] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0bd3b1e-bed8-467f-9a02-c57ebc06970c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.035818] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c748c1d2-c41f-45e6-bed5-e679d30ae56c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.040190] env[61985]: DEBUG nova.compute.manager [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received event network-vif-plugged-c4f26df9-9705-48e3-8414-595fed68fdee {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1065.040428] env[61985]: DEBUG oslo_concurrency.lockutils [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] Acquiring lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.040651] env[61985]: DEBUG oslo_concurrency.lockutils [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.040829] env[61985]: DEBUG oslo_concurrency.lockutils [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.041035] env[61985]: DEBUG nova.compute.manager [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] No waiting events found dispatching network-vif-plugged-c4f26df9-9705-48e3-8414-595fed68fdee {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1065.041227] env[61985]: WARNING nova.compute.manager [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received unexpected event network-vif-plugged-c4f26df9-9705-48e3-8414-595fed68fdee for instance with vm_state building and task_state spawning. [ 1065.041393] env[61985]: DEBUG nova.compute.manager [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received event network-changed-c4f26df9-9705-48e3-8414-595fed68fdee {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1065.041554] env[61985]: DEBUG nova.compute.manager [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Refreshing instance network info cache due to event network-changed-c4f26df9-9705-48e3-8414-595fed68fdee. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1065.041745] env[61985]: DEBUG oslo_concurrency.lockutils [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] Acquiring lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.041927] env[61985]: DEBUG oslo_concurrency.lockutils [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] Acquired lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.042069] env[61985]: DEBUG nova.network.neutron [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Refreshing network info cache for port c4f26df9-9705-48e3-8414-595fed68fdee {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1065.054618] env[61985]: INFO nova.compute.manager [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Took 20.53 seconds to build instance. [ 1065.056805] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252c26b7-6dce-4558-a5cc-19ff3f47319b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.075986] env[61985]: DEBUG nova.compute.provider_tree [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.135838] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936430, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.168172] env[61985]: DEBUG oslo_concurrency.lockutils [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.168360] env[61985]: DEBUG oslo_concurrency.lockutils [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.169299] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1123ef29-b457-4691-afbd-1827858f9ec6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.189696] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79141d50-bbe3-4e26-9f2d-04df726e6fa3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.218204] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Reconfiguring VM to detach interface {{(pid=61985) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1065.218514] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3413e9da-7e03-462a-aa28-2efb0755b6d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.239353] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1065.239353] env[61985]: value = "task-936431" [ 1065.239353] env[61985]: _type = "Task" [ 1065.239353] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.247555] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.340344] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1065.340877] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-940d85ee-be14-41b2-849b-7c4b8b98a957 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.350099] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1065.350099] env[61985]: value = "task-936432" [ 1065.350099] env[61985]: _type = "Task" [ 1065.350099] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.361840] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936432, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.453917] env[61985]: DEBUG nova.network.neutron [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Successfully updated port: 5a9c3e06-ff99-444a-a9ed-8f508e1854fb {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1065.552662] env[61985]: DEBUG oslo_concurrency.lockutils [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.566019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d8f9a822-7aaa-4e70-9ca5-d057fb0b70b7 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.056s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.566019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.011s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.566019] env[61985]: DEBUG nova.compute.manager [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.566019] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b6bb34-ff1e-45d1-8439-2588653348c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.572302] env[61985]: DEBUG nova.compute.manager [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61985) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3374}} [ 1065.573474] env[61985]: DEBUG nova.objects.instance [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lazy-loading 'flavor' on Instance uuid 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.579533] env[61985]: DEBUG nova.scheduler.client.report [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.589110] env[61985]: DEBUG nova.network.neutron [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1065.637174] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936430, 'name': ReconfigVM_Task, 'duration_secs': 0.854909} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.637174] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 036fae78-2af0-49f1-85fa-245c29f35ebc/036fae78-2af0-49f1-85fa-245c29f35ebc.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.637895] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-44f95c1f-87c5-452b-9fea-2105ff30b235 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.648037] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1065.648037] env[61985]: value = "task-936433" [ 1065.648037] env[61985]: _type = "Task" [ 1065.648037] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.663548] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936433, 'name': Rename_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.680533] env[61985]: DEBUG nova.network.neutron [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.750501] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.859358] env[61985]: DEBUG oslo_vmware.api [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936432, 'name': PowerOnVM_Task, 'duration_secs': 0.390347} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.859906] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1065.860360] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8067b8f1-9a3f-4a76-8ad9-bb373f7ccbf4 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance '8c4a1fe4-4ded-4064-8261-24826181a68c' progress to 100 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1065.953790] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "refresh_cache-decfe088-90d6-49f9-9eae-7aa461219eaf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.954153] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "refresh_cache-decfe088-90d6-49f9-9eae-7aa461219eaf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.955341] env[61985]: DEBUG nova.network.neutron [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1066.080585] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1066.080863] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd60cea6-a064-4728-8804-1616f52c11f5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.085036] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.393s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.086791] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.455s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.087074] env[61985]: DEBUG nova.objects.instance [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lazy-loading 'resources' on Instance uuid 5cd5e044-b0da-4564-8c2c-e894eb29a74c {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.089777] env[61985]: DEBUG oslo_vmware.api [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1066.089777] env[61985]: value = "task-936434" [ 1066.089777] env[61985]: _type = "Task" [ 1066.089777] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.104022] env[61985]: DEBUG oslo_vmware.api [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936434, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.160526] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936433, 'name': Rename_Task, 'duration_secs': 0.225019} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.161459] env[61985]: INFO nova.network.neutron [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updating port 47c6b710-9517-4551-b70b-d11e6cdd7c8f with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1066.163925] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1066.163992] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fbe83fd2-b4c5-4d6c-9e12-4fa18078c058 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.171006] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1066.171006] env[61985]: value = "task-936435" [ 1066.171006] env[61985]: _type = "Task" [ 1066.171006] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.182896] env[61985]: DEBUG oslo_concurrency.lockutils [req-771458e3-5cc4-4902-95d5-861b4cf613b3 req-deff9119-cf0b-4167-9544-b8397075e222 service nova] Releasing lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.183325] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936435, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.254276] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.498562] env[61985]: DEBUG nova.network.neutron [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Successfully updated port: 6a2e1f2a-6455-4354-83df-19ac0b8f17db {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1066.540229] env[61985]: DEBUG nova.network.neutron [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1066.602447] env[61985]: DEBUG oslo_vmware.api [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936434, 'name': PowerOffVM_Task, 'duration_secs': 0.23622} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.602724] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1066.602888] env[61985]: DEBUG nova.compute.manager [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1066.603671] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882c2d7b-2e4a-4d8d-87e7-c80d06994fac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.681723] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936435, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.754867] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.844661] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dfb4533-7341-4e3b-8a6f-758a9e3390b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.855023] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2e67b4-5665-4b21-8449-3e4503cc518d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.887634] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259c824a-b48f-41ee-af31-313477e6ef0e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.895144] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f718b51c-75db-41ed-a980-96d11a247de8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.908069] env[61985]: DEBUG nova.compute.provider_tree [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.974913] env[61985]: DEBUG nova.network.neutron [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Updating instance_info_cache with network_info: [{"id": "5a9c3e06-ff99-444a-a9ed-8f508e1854fb", "address": "fa:16:3e:14:41:8e", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a9c3e06-ff", "ovs_interfaceid": "5a9c3e06-ff99-444a-a9ed-8f508e1854fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.999933] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.000304] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquired lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.000304] env[61985]: DEBUG nova.network.neutron [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1067.063560] env[61985]: DEBUG nova.compute.manager [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Received event network-vif-plugged-5a9c3e06-ff99-444a-a9ed-8f508e1854fb {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1067.063805] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Acquiring lock "decfe088-90d6-49f9-9eae-7aa461219eaf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.064273] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Lock "decfe088-90d6-49f9-9eae-7aa461219eaf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.064577] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Lock "decfe088-90d6-49f9-9eae-7aa461219eaf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.064807] env[61985]: DEBUG nova.compute.manager [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] No waiting events found dispatching network-vif-plugged-5a9c3e06-ff99-444a-a9ed-8f508e1854fb {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1067.065050] env[61985]: WARNING nova.compute.manager [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Received unexpected event network-vif-plugged-5a9c3e06-ff99-444a-a9ed-8f508e1854fb for instance with vm_state building and task_state spawning. [ 1067.065542] env[61985]: DEBUG nova.compute.manager [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Received event network-changed-5a9c3e06-ff99-444a-a9ed-8f508e1854fb {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1067.065776] env[61985]: DEBUG nova.compute.manager [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Refreshing instance network info cache due to event network-changed-5a9c3e06-ff99-444a-a9ed-8f508e1854fb. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1067.065999] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Acquiring lock "refresh_cache-decfe088-90d6-49f9-9eae-7aa461219eaf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.117748] env[61985]: DEBUG oslo_concurrency.lockutils [None req-19a6d47e-4da8-4af2-8761-cee1ae365d84 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.554s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.182823] env[61985]: DEBUG oslo_vmware.api [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936435, 'name': PowerOnVM_Task, 'duration_secs': 0.557761} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.183101] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1067.183333] env[61985]: INFO nova.compute.manager [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Took 9.00 seconds to spawn the instance on the hypervisor. [ 1067.183528] env[61985]: DEBUG nova.compute.manager [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1067.184339] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6044115-2b57-4bec-bacf-cf63f49ba61e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.253910] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.411108] env[61985]: DEBUG nova.scheduler.client.report [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.477509] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "refresh_cache-decfe088-90d6-49f9-9eae-7aa461219eaf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.477949] env[61985]: DEBUG nova.compute.manager [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Instance network_info: |[{"id": "5a9c3e06-ff99-444a-a9ed-8f508e1854fb", "address": "fa:16:3e:14:41:8e", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a9c3e06-ff", "ovs_interfaceid": "5a9c3e06-ff99-444a-a9ed-8f508e1854fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1067.478288] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Acquired lock "refresh_cache-decfe088-90d6-49f9-9eae-7aa461219eaf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.478481] env[61985]: DEBUG nova.network.neutron [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Refreshing network info cache for port 5a9c3e06-ff99-444a-a9ed-8f508e1854fb {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1067.480975] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:41:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55520f67-d092-4eb7-940f-d7cceaa1ca1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5a9c3e06-ff99-444a-a9ed-8f508e1854fb', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1067.489161] env[61985]: DEBUG oslo.service.loopingcall [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.490232] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1067.490602] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e4fca2b-506f-4ef5-81ac-108cd844ac01 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.514611] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1067.514611] env[61985]: value = "task-936436" [ 1067.514611] env[61985]: _type = "Task" [ 1067.514611] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.521546] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936436, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.542632] env[61985]: DEBUG nova.network.neutron [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1067.633409] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.633638] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.703069] env[61985]: INFO nova.compute.manager [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Took 21.93 seconds to build instance. [ 1067.755897] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.917561] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.831s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.922741] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.014s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.924617] env[61985]: INFO nova.compute.claims [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1067.951120] env[61985]: INFO nova.scheduler.client.report [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Deleted allocations for instance 5cd5e044-b0da-4564-8c2c-e894eb29a74c [ 1068.023014] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936436, 'name': CreateVM_Task, 'duration_secs': 0.428175} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.023222] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1068.023933] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.024128] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.024452] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1068.024717] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f12c4d93-73e3-4cca-b488-a410a9bd806d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.035362] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1068.035362] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52482b83-40f1-843d-b9e2-fa0f327141ec" [ 1068.035362] env[61985]: _type = "Task" [ 1068.035362] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.040953] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52482b83-40f1-843d-b9e2-fa0f327141ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.145695] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1068.146079] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1068.205222] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2522a05-0a3b-4dc2-8e0b-448b4bdf065f tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.441s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.260510] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.324161] env[61985]: DEBUG nova.network.neutron [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Updated VIF entry in instance network info cache for port 5a9c3e06-ff99-444a-a9ed-8f508e1854fb. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1068.324161] env[61985]: DEBUG nova.network.neutron [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Updating instance_info_cache with network_info: [{"id": "5a9c3e06-ff99-444a-a9ed-8f508e1854fb", "address": "fa:16:3e:14:41:8e", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a9c3e06-ff", "ovs_interfaceid": "5a9c3e06-ff99-444a-a9ed-8f508e1854fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.382629] env[61985]: DEBUG nova.network.neutron [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Updating instance_info_cache with network_info: [{"id": "01e88a1a-79a9-4c6f-a939-607ae4ecd7f2", "address": "fa:16:3e:08:9d:22", "network": {"id": "94f9d52d-2ccd-44f4-85f0-c6c953a38dec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1611261022", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.202", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01e88a1a-79", "ovs_interfaceid": "01e88a1a-79a9-4c6f-a939-607ae4ecd7f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c4f26df9-9705-48e3-8414-595fed68fdee", "address": "fa:16:3e:03:51:98", "network": {"id": "53d5d450-bde4-49ea-a478-f33c50d1c841", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-748247095", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.251", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4f26df9-97", "ovs_interfaceid": "c4f26df9-9705-48e3-8414-595fed68fdee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a2e1f2a-6455-4354-83df-19ac0b8f17db", "address": "fa:16:3e:78:f7:aa", "network": {"id": "94f9d52d-2ccd-44f4-85f0-c6c953a38dec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1611261022", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.174", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a2e1f2a-64", "ovs_interfaceid": "6a2e1f2a-6455-4354-83df-19ac0b8f17db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.460769] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bd4bc79f-a4af-496b-bbfd-2162d9b13dc8 tempest-ImagesTestJSON-959827617 tempest-ImagesTestJSON-959827617-project-member] Lock "5cd5e044-b0da-4564-8c2c-e894eb29a74c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.041s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.544558] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52482b83-40f1-843d-b9e2-fa0f327141ec, 'name': SearchDatastore_Task, 'duration_secs': 0.00991} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.545258] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.545692] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1068.546252] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.546616] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.547810] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1068.549169] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-60f2f9c2-e6a3-4504-9fe3-3171425407d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.559586] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1068.559586] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1068.559586] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55c3cbe4-d843-4979-95b3-a7cb89414671 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.564827] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1068.564827] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520e9147-2da7-fea4-9c4b-303d03a12cda" [ 1068.564827] env[61985]: _type = "Task" [ 1068.564827] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.573849] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520e9147-2da7-fea4-9c4b-303d03a12cda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.761706] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.827522] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Releasing lock "refresh_cache-decfe088-90d6-49f9-9eae-7aa461219eaf" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.827522] env[61985]: DEBUG nova.compute.manager [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received event network-vif-plugged-6a2e1f2a-6455-4354-83df-19ac0b8f17db {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1068.827522] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Acquiring lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.827522] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.827522] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.827522] env[61985]: DEBUG nova.compute.manager [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] No waiting events found dispatching network-vif-plugged-6a2e1f2a-6455-4354-83df-19ac0b8f17db {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1068.827522] env[61985]: WARNING nova.compute.manager [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received unexpected event network-vif-plugged-6a2e1f2a-6455-4354-83df-19ac0b8f17db for instance with vm_state building and task_state spawning. [ 1068.827522] env[61985]: DEBUG nova.compute.manager [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received event network-changed-6a2e1f2a-6455-4354-83df-19ac0b8f17db {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1068.827522] env[61985]: DEBUG nova.compute.manager [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Refreshing instance network info cache due to event network-changed-6a2e1f2a-6455-4354-83df-19ac0b8f17db. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1068.827522] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Acquiring lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.886019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.886019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.886019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.886019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.886019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.886019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Releasing lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.887313] env[61985]: DEBUG nova.compute.manager [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Instance network_info: |[{"id": "01e88a1a-79a9-4c6f-a939-607ae4ecd7f2", "address": "fa:16:3e:08:9d:22", "network": {"id": "94f9d52d-2ccd-44f4-85f0-c6c953a38dec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1611261022", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.202", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01e88a1a-79", "ovs_interfaceid": "01e88a1a-79a9-4c6f-a939-607ae4ecd7f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c4f26df9-9705-48e3-8414-595fed68fdee", "address": "fa:16:3e:03:51:98", "network": {"id": "53d5d450-bde4-49ea-a478-f33c50d1c841", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-748247095", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.251", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4f26df9-97", "ovs_interfaceid": "c4f26df9-9705-48e3-8414-595fed68fdee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a2e1f2a-6455-4354-83df-19ac0b8f17db", "address": "fa:16:3e:78:f7:aa", "network": {"id": "94f9d52d-2ccd-44f4-85f0-c6c953a38dec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1611261022", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.174", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a2e1f2a-64", "ovs_interfaceid": "6a2e1f2a-6455-4354-83df-19ac0b8f17db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1068.889535] env[61985]: INFO nova.compute.manager [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Terminating instance [ 1068.892016] env[61985]: DEBUG nova.compute.manager [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1068.892829] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1068.893209] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Acquired lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.893539] env[61985]: DEBUG nova.network.neutron [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Refreshing network info cache for port 6a2e1f2a-6455-4354-83df-19ac0b8f17db {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1068.897143] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:9d:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea4a9e02-45f1-4afb-8abb-0de26b153086', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01e88a1a-79a9-4c6f-a939-607ae4ecd7f2', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:51:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4734e5e-2a76-4bda-8905-70c9bf9e007f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4f26df9-9705-48e3-8414-595fed68fdee', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:f7:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea4a9e02-45f1-4afb-8abb-0de26b153086', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6a2e1f2a-6455-4354-83df-19ac0b8f17db', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1068.906930] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Creating folder: Project (60ecaef69401418eac6c655c1b7df5be). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1068.907769] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3bf402-4ab7-4bd5-b1a4-e38c259e34fd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.911065] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-79921637-b113-45a9-b3f7-852513a5a8c2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.918407] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1068.919306] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f0ba2a9-cfb9-4fab-8889-8e49a2b160bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.922213] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Created folder: Project (60ecaef69401418eac6c655c1b7df5be) in parent group-v211285. [ 1068.922985] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Creating folder: Instances. Parent ref: group-v211459. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1068.922985] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-89623eb2-2ae1-4761-83df-734425db610f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.942543] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Created folder: Instances in parent group-v211459. [ 1068.942543] env[61985]: DEBUG oslo.service.loopingcall [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.942543] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1068.942543] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67403283-78f0-462c-8d07-e6c3e9f6a4af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.976806] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1068.976806] env[61985]: value = "task-936440" [ 1068.976806] env[61985]: _type = "Task" [ 1068.976806] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.987868] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936440, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.009235] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1069.009784] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1069.010047] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleting the datastore file [datastore1] 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.010545] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a38753c-b7bd-4b3c-bc3a-3595ea7391a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.017373] env[61985]: DEBUG oslo_vmware.api [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1069.017373] env[61985]: value = "task-936441" [ 1069.017373] env[61985]: _type = "Task" [ 1069.017373] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.028498] env[61985]: DEBUG oslo_vmware.api [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936441, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.078148] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520e9147-2da7-fea4-9c4b-303d03a12cda, 'name': SearchDatastore_Task, 'duration_secs': 0.008297} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.082875] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-781783f4-9c76-4db3-a531-0fdfdfae235f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.091074] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1069.091074] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52164e01-a55f-14d0-0567-f7ef19a753d3" [ 1069.091074] env[61985]: _type = "Task" [ 1069.091074] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.097128] env[61985]: DEBUG nova.compute.manager [req-6bdc8cf4-c2a8-4701-b570-a6bf1e36ae6f req-41e15eb3-f55e-4902-882a-adb8bd6d9bb1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Received event network-vif-plugged-47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1069.097412] env[61985]: DEBUG oslo_concurrency.lockutils [req-6bdc8cf4-c2a8-4701-b570-a6bf1e36ae6f req-41e15eb3-f55e-4902-882a-adb8bd6d9bb1 service nova] Acquiring lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.097908] env[61985]: DEBUG oslo_concurrency.lockutils [req-6bdc8cf4-c2a8-4701-b570-a6bf1e36ae6f req-41e15eb3-f55e-4902-882a-adb8bd6d9bb1 service nova] Lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.097908] env[61985]: DEBUG oslo_concurrency.lockutils [req-6bdc8cf4-c2a8-4701-b570-a6bf1e36ae6f req-41e15eb3-f55e-4902-882a-adb8bd6d9bb1 service nova] Lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.098051] env[61985]: DEBUG nova.compute.manager [req-6bdc8cf4-c2a8-4701-b570-a6bf1e36ae6f req-41e15eb3-f55e-4902-882a-adb8bd6d9bb1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] No waiting events found dispatching network-vif-plugged-47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1069.099562] env[61985]: WARNING nova.compute.manager [req-6bdc8cf4-c2a8-4701-b570-a6bf1e36ae6f req-41e15eb3-f55e-4902-882a-adb8bd6d9bb1 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Received unexpected event network-vif-plugged-47c6b710-9517-4551-b70b-d11e6cdd7c8f for instance with vm_state shelved_offloaded and task_state spawning. [ 1069.113255] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52164e01-a55f-14d0-0567-f7ef19a753d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.142434] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.142434] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.142781] env[61985]: DEBUG nova.network.neutron [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1069.157196] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "8c4a1fe4-4ded-4064-8261-24826181a68c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.157460] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.157671] env[61985]: DEBUG nova.compute.manager [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Going to confirm migration 5 {{(pid=61985) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4789}} [ 1069.245345] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bdb8856-ee26-4e7f-aef1-79c23e39ad0e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.264430] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313a8f4b-42c8-4b17-8cb8-6645330934ae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.269999] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.301968] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b970d254-88ff-44f2-aa7f-d48bd68cb238 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.309573] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b157d7-d1db-4893-9c78-a2a916f24537 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.325107] env[61985]: DEBUG nova.compute.provider_tree [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.488499] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936440, 'name': CreateVM_Task, 'duration_secs': 0.43289} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.488691] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1069.489601] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.489774] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.490126] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1069.490390] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d12ba687-ec21-4249-b4d1-fa4047838f9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.494919] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1069.494919] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520a9d4a-911e-89fc-f4d8-f6c8071c637a" [ 1069.494919] env[61985]: _type = "Task" [ 1069.494919] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.502989] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520a9d4a-911e-89fc-f4d8-f6c8071c637a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.525921] env[61985]: DEBUG oslo_vmware.api [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936441, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201661} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.525921] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.526105] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1069.526286] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1069.526361] env[61985]: INFO nova.compute.manager [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Took 0.63 seconds to destroy the instance on the hypervisor. [ 1069.526547] env[61985]: DEBUG oslo.service.loopingcall [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.526738] env[61985]: DEBUG nova.compute.manager [-] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1069.526834] env[61985]: DEBUG nova.network.neutron [-] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1069.600195] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52164e01-a55f-14d0-0567-f7ef19a753d3, 'name': SearchDatastore_Task, 'duration_secs': 0.010501} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.602408] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.602681] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] decfe088-90d6-49f9-9eae-7aa461219eaf/decfe088-90d6-49f9-9eae-7aa461219eaf.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1069.602951] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eee85880-c135-4ee3-bacc-bf0136d11152 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.610105] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1069.610105] env[61985]: value = "task-936442" [ 1069.610105] env[61985]: _type = "Task" [ 1069.610105] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.619984] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936442, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.762260] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.763749] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.764040] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.764383] env[61985]: DEBUG nova.network.neutron [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1069.764664] env[61985]: DEBUG nova.objects.instance [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lazy-loading 'info_cache' on Instance uuid 8c4a1fe4-4ded-4064-8261-24826181a68c {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.828080] env[61985]: DEBUG nova.scheduler.client.report [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1069.928525] env[61985]: DEBUG nova.compute.manager [req-52ac7ba3-8af3-46cc-82eb-f018762eab72 req-31d47ced-a19e-4005-a1d4-84addd487ed7 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Received event network-changed-52d657fa-a084-485b-858a-368bb739f4fc {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1069.928525] env[61985]: DEBUG nova.compute.manager [req-52ac7ba3-8af3-46cc-82eb-f018762eab72 req-31d47ced-a19e-4005-a1d4-84addd487ed7 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Refreshing instance network info cache due to event network-changed-52d657fa-a084-485b-858a-368bb739f4fc. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1069.928525] env[61985]: DEBUG oslo_concurrency.lockutils [req-52ac7ba3-8af3-46cc-82eb-f018762eab72 req-31d47ced-a19e-4005-a1d4-84addd487ed7 service nova] Acquiring lock "refresh_cache-036fae78-2af0-49f1-85fa-245c29f35ebc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.928525] env[61985]: DEBUG oslo_concurrency.lockutils [req-52ac7ba3-8af3-46cc-82eb-f018762eab72 req-31d47ced-a19e-4005-a1d4-84addd487ed7 service nova] Acquired lock "refresh_cache-036fae78-2af0-49f1-85fa-245c29f35ebc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.928525] env[61985]: DEBUG nova.network.neutron [req-52ac7ba3-8af3-46cc-82eb-f018762eab72 req-31d47ced-a19e-4005-a1d4-84addd487ed7 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Refreshing network info cache for port 52d657fa-a084-485b-858a-368bb739f4fc {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1070.007745] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520a9d4a-911e-89fc-f4d8-f6c8071c637a, 'name': SearchDatastore_Task, 'duration_secs': 0.011853} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.008231] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.008514] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.008825] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.009044] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.009386] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.009710] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-834da7fa-f24e-4851-8efd-7c06bbb1950d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.020439] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.020682] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1070.021478] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dccab31b-f3a5-4cf3-a6d1-d99c4e34c15b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.027022] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1070.027022] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5214e16d-7013-0985-eebf-2fc5f89c7efb" [ 1070.027022] env[61985]: _type = "Task" [ 1070.027022] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.031718] env[61985]: DEBUG nova.network.neutron [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Updated VIF entry in instance network info cache for port 6a2e1f2a-6455-4354-83df-19ac0b8f17db. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1070.031971] env[61985]: DEBUG nova.network.neutron [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Updating instance_info_cache with network_info: [{"id": "01e88a1a-79a9-4c6f-a939-607ae4ecd7f2", "address": "fa:16:3e:08:9d:22", "network": {"id": "94f9d52d-2ccd-44f4-85f0-c6c953a38dec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1611261022", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.202", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01e88a1a-79", "ovs_interfaceid": "01e88a1a-79a9-4c6f-a939-607ae4ecd7f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c4f26df9-9705-48e3-8414-595fed68fdee", "address": "fa:16:3e:03:51:98", "network": {"id": "53d5d450-bde4-49ea-a478-f33c50d1c841", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-748247095", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.251", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4734e5e-2a76-4bda-8905-70c9bf9e007f", "external-id": "nsx-vlan-transportzone-122", "segmentation_id": 122, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4f26df9-97", "ovs_interfaceid": "c4f26df9-9705-48e3-8414-595fed68fdee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a2e1f2a-6455-4354-83df-19ac0b8f17db", "address": "fa:16:3e:78:f7:aa", "network": {"id": "94f9d52d-2ccd-44f4-85f0-c6c953a38dec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1611261022", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.174", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a2e1f2a-64", "ovs_interfaceid": "6a2e1f2a-6455-4354-83df-19ac0b8f17db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.041957] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5214e16d-7013-0985-eebf-2fc5f89c7efb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.123724] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936442, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.246365] env[61985]: DEBUG nova.network.neutron [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updating instance_info_cache with network_info: [{"id": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "address": "fa:16:3e:af:c8:fd", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47c6b710-95", "ovs_interfaceid": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.259376] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.336019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.336019] env[61985]: DEBUG nova.compute.manager [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1070.534771] env[61985]: DEBUG oslo_concurrency.lockutils [req-630e9288-763a-4ca3-b3a4-058dea7d83cf req-d772be11-bfaf-42b0-bdeb-7dd8a63fe7d1 service nova] Releasing lock "refresh_cache-8eea60eb-93cf-4701-9ba3-0805691ca016" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.542267] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5214e16d-7013-0985-eebf-2fc5f89c7efb, 'name': SearchDatastore_Task, 'duration_secs': 0.011746} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.543186] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6ada8fa-3038-4188-94ce-3faa61c9a947 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.554022] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1070.554022] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521dbc90-e5bb-4211-8334-32161650f945" [ 1070.554022] env[61985]: _type = "Task" [ 1070.554022] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.561601] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521dbc90-e5bb-4211-8334-32161650f945, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.593963] env[61985]: DEBUG nova.network.neutron [-] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.629233] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936442, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.778855} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.629233] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] decfe088-90d6-49f9-9eae-7aa461219eaf/decfe088-90d6-49f9-9eae-7aa461219eaf.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1070.629377] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1070.629560] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-46e1a8f9-e156-41fe-a518-e613c3313c28 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.642022] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1070.642022] env[61985]: value = "task-936443" [ 1070.642022] env[61985]: _type = "Task" [ 1070.642022] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.647953] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936443, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.748924] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.768495] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.791114] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='70436504cf8927cdafb0ebf9e6d85e62',container_format='bare',created_at=2024-09-18T01:31:52Z,direct_url=,disk_format='vmdk',id=173db284-6b8b-4355-bdde-6a9ba3be888e,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-347683918-shelved',owner='be4f1b894ce04f37a0d106e93f4101fa',properties=ImageMetaProps,protected=,size=31665152,status='active',tags=,updated_at=2024-09-18T01:32:09Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1070.791114] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1070.791114] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.791114] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1070.791114] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.791533] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1070.791924] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1070.792304] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1070.792658] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1070.792995] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1070.793350] env[61985]: DEBUG nova.virt.hardware [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1070.794348] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb72cf37-8b3c-456f-a9fa-c2354a5870c2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.803537] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6079bc4a-4a5f-43c0-b46e-157c3be83dcc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.824021] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:c8:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cb0e556a-0f69-4a5c-af62-ffc46edb8e63', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47c6b710-9517-4551-b70b-d11e6cdd7c8f', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1070.830312] env[61985]: DEBUG oslo.service.loopingcall [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.833361] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1070.833717] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba111cb2-8349-4da7-9b3d-15a0ee31c6fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.854024] env[61985]: DEBUG nova.compute.utils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1070.854024] env[61985]: DEBUG nova.compute.manager [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1070.854024] env[61985]: DEBUG nova.network.neutron [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1070.860249] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1070.860249] env[61985]: value = "task-936444" [ 1070.860249] env[61985]: _type = "Task" [ 1070.860249] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.868641] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936444, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.067772] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521dbc90-e5bb-4211-8334-32161650f945, 'name': SearchDatastore_Task, 'duration_secs': 0.031897} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.067991] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.068482] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 8eea60eb-93cf-4701-9ba3-0805691ca016/8eea60eb-93cf-4701-9ba3-0805691ca016.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1071.068619] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-00329021-0910-4d3c-9319-b72783dbbc03 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.077725] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1071.077725] env[61985]: value = "task-936445" [ 1071.077725] env[61985]: _type = "Task" [ 1071.077725] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.086451] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936445, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.097484] env[61985]: INFO nova.compute.manager [-] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Took 1.57 seconds to deallocate network for instance. [ 1071.147020] env[61985]: DEBUG nova.policy [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4f5b5ce76e8b4fae9c905bc7a39b683e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4923009773c047519b883745c59e3516', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1071.152150] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936443, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070743} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.152412] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1071.153566] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7053d49-a694-45bb-bb9d-f78050e986ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.175846] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] decfe088-90d6-49f9-9eae-7aa461219eaf/decfe088-90d6-49f9-9eae-7aa461219eaf.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.177704] env[61985]: DEBUG nova.network.neutron [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance_info_cache with network_info: [{"id": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "address": "fa:16:3e:35:b3:a2", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88467e26-c1", "ovs_interfaceid": "88467e26-c1c7-4b6a-ab2b-9d86b552c0c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.180399] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5a80dee-df0b-4f27-9c4e-5f87424e067d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.203967] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1071.203967] env[61985]: value = "task-936446" [ 1071.203967] env[61985]: _type = "Task" [ 1071.203967] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.212964] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936446, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.248541] env[61985]: DEBUG nova.compute.manager [req-328d70ba-a455-4362-b7f1-5c25b2a0478a req-3d231eec-d1b5-4aee-a4f4-9a7033f3c870 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Received event network-changed-47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1071.248863] env[61985]: DEBUG nova.compute.manager [req-328d70ba-a455-4362-b7f1-5c25b2a0478a req-3d231eec-d1b5-4aee-a4f4-9a7033f3c870 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Refreshing instance network info cache due to event network-changed-47c6b710-9517-4551-b70b-d11e6cdd7c8f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1071.249607] env[61985]: DEBUG oslo_concurrency.lockutils [req-328d70ba-a455-4362-b7f1-5c25b2a0478a req-3d231eec-d1b5-4aee-a4f4-9a7033f3c870 service nova] Acquiring lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.249854] env[61985]: DEBUG oslo_concurrency.lockutils [req-328d70ba-a455-4362-b7f1-5c25b2a0478a req-3d231eec-d1b5-4aee-a4f4-9a7033f3c870 service nova] Acquired lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.250127] env[61985]: DEBUG nova.network.neutron [req-328d70ba-a455-4362-b7f1-5c25b2a0478a req-3d231eec-d1b5-4aee-a4f4-9a7033f3c870 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Refreshing network info cache for port 47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1071.265761] env[61985]: DEBUG oslo_vmware.api [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936431, 'name': ReconfigVM_Task, 'duration_secs': 5.908341} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.267235] env[61985]: DEBUG oslo_concurrency.lockutils [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.267756] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Reconfigured VM to detach interface {{(pid=61985) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1071.343215] env[61985]: DEBUG oslo_concurrency.lockutils [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "885a36d9-24c6-407a-9d6b-01cd97e90674" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.343550] env[61985]: DEBUG oslo_concurrency.lockutils [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.356214] env[61985]: DEBUG nova.compute.manager [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1071.382118] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936444, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.385766] env[61985]: DEBUG nova.network.neutron [req-52ac7ba3-8af3-46cc-82eb-f018762eab72 req-31d47ced-a19e-4005-a1d4-84addd487ed7 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Updated VIF entry in instance network info cache for port 52d657fa-a084-485b-858a-368bb739f4fc. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1071.386188] env[61985]: DEBUG nova.network.neutron [req-52ac7ba3-8af3-46cc-82eb-f018762eab72 req-31d47ced-a19e-4005-a1d4-84addd487ed7 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Updating instance_info_cache with network_info: [{"id": "52d657fa-a084-485b-858a-368bb739f4fc", "address": "fa:16:3e:8f:ab:a9", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52d657fa-a0", "ovs_interfaceid": "52d657fa-a084-485b-858a-368bb739f4fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.590150] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936445, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.604757] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.604938] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.605361] env[61985]: DEBUG nova.objects.instance [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lazy-loading 'resources' on Instance uuid 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.680158] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-8c4a1fe4-4ded-4064-8261-24826181a68c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.680417] env[61985]: DEBUG nova.objects.instance [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lazy-loading 'migration_context' on Instance uuid 8c4a1fe4-4ded-4064-8261-24826181a68c {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.716542] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936446, 'name': ReconfigVM_Task, 'duration_secs': 0.267471} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.716542] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Reconfigured VM instance instance-00000061 to attach disk [datastore1] decfe088-90d6-49f9-9eae-7aa461219eaf/decfe088-90d6-49f9-9eae-7aa461219eaf.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.716542] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-409b70cd-6b1c-4a56-8a0e-37b27ccc066c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.728075] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1071.728075] env[61985]: value = "task-936447" [ 1071.728075] env[61985]: _type = "Task" [ 1071.728075] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.740600] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936447, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.848260] env[61985]: DEBUG nova.compute.utils [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1071.877634] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936444, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.893325] env[61985]: DEBUG oslo_concurrency.lockutils [req-52ac7ba3-8af3-46cc-82eb-f018762eab72 req-31d47ced-a19e-4005-a1d4-84addd487ed7 service nova] Releasing lock "refresh_cache-036fae78-2af0-49f1-85fa-245c29f35ebc" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.992389] env[61985]: DEBUG nova.network.neutron [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Successfully created port: 1a081dc2-f497-4708-a3af-0b373d04cb85 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1072.076400] env[61985]: DEBUG nova.compute.manager [req-5fa05411-7ff6-476c-b8d9-f906b4e59244 req-ac85574d-4e7a-400d-999f-5296b8032a10 service nova] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Received event network-vif-deleted-96ab3107-6606-4dd8-9b69-7d2445a81994 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1072.091814] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936445, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.796493} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.092158] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 8eea60eb-93cf-4701-9ba3-0805691ca016/8eea60eb-93cf-4701-9ba3-0805691ca016.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1072.092265] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.092536] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae889abf-92cd-4fb0-a3a5-1407867de98e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.105045] env[61985]: DEBUG nova.network.neutron [req-328d70ba-a455-4362-b7f1-5c25b2a0478a req-3d231eec-d1b5-4aee-a4f4-9a7033f3c870 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updated VIF entry in instance network info cache for port 47c6b710-9517-4551-b70b-d11e6cdd7c8f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1072.105045] env[61985]: DEBUG nova.network.neutron [req-328d70ba-a455-4362-b7f1-5c25b2a0478a req-3d231eec-d1b5-4aee-a4f4-9a7033f3c870 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updating instance_info_cache with network_info: [{"id": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "address": "fa:16:3e:af:c8:fd", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47c6b710-95", "ovs_interfaceid": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.106715] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1072.106715] env[61985]: value = "task-936448" [ 1072.106715] env[61985]: _type = "Task" [ 1072.106715] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.125646] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936448, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.186347] env[61985]: DEBUG nova.objects.base [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Object Instance<8c4a1fe4-4ded-4064-8261-24826181a68c> lazy-loaded attributes: info_cache,migration_context {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1072.187427] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36148670-847e-43c8-9210-48a7866ed56d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.217823] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb4cc335-3b72-4eca-ae67-856d36af1da0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.234019] env[61985]: DEBUG oslo_vmware.api [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1072.234019] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528c8932-dfc4-7462-6acd-eee74cb1e266" [ 1072.234019] env[61985]: _type = "Task" [ 1072.234019] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.251570] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936447, 'name': Rename_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.255722] env[61985]: DEBUG oslo_vmware.api [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528c8932-dfc4-7462-6acd-eee74cb1e266, 'name': SearchDatastore_Task, 'duration_secs': 0.010322} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.258276] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.351665] env[61985]: DEBUG oslo_concurrency.lockutils [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.367432] env[61985]: DEBUG nova.compute.manager [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1072.383664] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936444, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.396279] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1072.396945] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1072.396945] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1072.396945] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1072.397084] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1072.397190] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1072.397407] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1072.397574] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1072.397787] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1072.397965] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1072.398165] env[61985]: DEBUG nova.virt.hardware [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1072.399016] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade98678-0531-4722-aec3-e27fe33257f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.408106] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5981e65-bb1e-4d46-b657-fc8da8053e72 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.413178] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d48fb834-903f-48d6-80db-4985bfd07cd5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.429264] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef65148-f279-4521-abc6-aeecfe8929b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.464167] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6183be-b430-4034-852e-1ae8ddadec2d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.472325] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4911cd3b-5f44-46a9-bc57-e66f2475336b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.488261] env[61985]: DEBUG nova.compute.provider_tree [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.610082] env[61985]: DEBUG oslo_concurrency.lockutils [req-328d70ba-a455-4362-b7f1-5c25b2a0478a req-3d231eec-d1b5-4aee-a4f4-9a7033f3c870 service nova] Releasing lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.620147] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936448, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06722} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.620464] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1072.621314] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72dd6c20-cae9-4d36-8654-54460abf2cf7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.651089] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 8eea60eb-93cf-4701-9ba3-0805691ca016/8eea60eb-93cf-4701-9ba3-0805691ca016.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1072.651438] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6f7d691-ded1-4836-aafd-b43750739f27 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.666645] env[61985]: DEBUG oslo_concurrency.lockutils [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.666833] env[61985]: DEBUG oslo_concurrency.lockutils [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquired lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.667023] env[61985]: DEBUG nova.network.neutron [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1072.674532] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1072.674532] env[61985]: value = "task-936449" [ 1072.674532] env[61985]: _type = "Task" [ 1072.674532] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.689151] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936449, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.733157] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "8db82979-e527-4c63-8ce8-fe1456f7708c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.733487] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "8db82979-e527-4c63-8ce8-fe1456f7708c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.733775] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "8db82979-e527-4c63-8ce8-fe1456f7708c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.734033] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "8db82979-e527-4c63-8ce8-fe1456f7708c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.734296] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "8db82979-e527-4c63-8ce8-fe1456f7708c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.740043] env[61985]: INFO nova.compute.manager [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Terminating instance [ 1072.743080] env[61985]: DEBUG nova.compute.manager [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1072.743080] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1072.745013] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe1f0db-75eb-4d52-9903-f74b021312c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.751217] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936447, 'name': Rename_Task, 'duration_secs': 0.844069} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.751999] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1072.752276] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24263959-72ef-40b7-a93d-21a98ab541a3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.756898] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1072.757647] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6937d46c-ab24-49b7-9fae-15b0a1e41cad {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.762449] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1072.762449] env[61985]: value = "task-936450" [ 1072.762449] env[61985]: _type = "Task" [ 1072.762449] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.767418] env[61985]: DEBUG oslo_vmware.api [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1072.767418] env[61985]: value = "task-936451" [ 1072.767418] env[61985]: _type = "Task" [ 1072.767418] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.785899] env[61985]: DEBUG oslo_vmware.api [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936451, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.786900] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936450, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.885749] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936444, 'name': CreateVM_Task, 'duration_secs': 1.749916} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.886239] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1072.887357] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.887357] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "[datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.887357] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1072.887538] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccdfeee3-ae8e-448b-970c-955577d41a28 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.892599] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1072.892599] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52150799-b6c6-b59e-db69-f85defa06e56" [ 1072.892599] env[61985]: _type = "Task" [ 1072.892599] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.910859] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "[datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.911129] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Processing image 173db284-6b8b-4355-bdde-6a9ba3be888e {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1072.911393] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e/173db284-6b8b-4355-bdde-6a9ba3be888e.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.911546] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "[datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e/173db284-6b8b-4355-bdde-6a9ba3be888e.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.911728] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1072.912011] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff8a8d9f-3c2f-41a2-8780-e7d65402752b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.920847] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1072.921114] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1072.923117] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bcc9cf0-6e8f-48db-a6c0-eb34d2a1180f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.933177] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1072.933177] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52be941f-410a-cb01-2682-b82a4f3484dd" [ 1072.933177] env[61985]: _type = "Task" [ 1072.933177] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.944239] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52be941f-410a-cb01-2682-b82a4f3484dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.991645] env[61985]: DEBUG nova.scheduler.client.report [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1073.187123] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936449, 'name': ReconfigVM_Task, 'duration_secs': 0.456119} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.187513] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 8eea60eb-93cf-4701-9ba3-0805691ca016/8eea60eb-93cf-4701-9ba3-0805691ca016.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.188711] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5264d50b-ad14-4d46-b109-3332c9534231 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.195979] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1073.195979] env[61985]: value = "task-936452" [ 1073.195979] env[61985]: _type = "Task" [ 1073.195979] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.206828] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936452, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.261775] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquiring lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.262060] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.275677] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936450, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.281379] env[61985]: DEBUG oslo_vmware.api [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936451, 'name': PowerOffVM_Task, 'duration_secs': 0.291092} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.281792] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1073.281792] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1073.282308] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb8de54f-694e-48a5-8e68-3fed22532861 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.393234] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1073.395328] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1073.395328] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Deleting the datastore file [datastore2] 8db82979-e527-4c63-8ce8-fe1456f7708c {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.395328] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86c005cc-b8ee-4a50-a2d2-dafb9ab33177 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.400586] env[61985]: DEBUG oslo_vmware.api [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1073.400586] env[61985]: value = "task-936454" [ 1073.400586] env[61985]: _type = "Task" [ 1073.400586] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.408979] env[61985]: DEBUG oslo_vmware.api [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.429689] env[61985]: DEBUG oslo_concurrency.lockutils [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "885a36d9-24c6-407a-9d6b-01cd97e90674" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.429994] env[61985]: DEBUG oslo_concurrency.lockutils [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.430282] env[61985]: INFO nova.compute.manager [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Attaching volume 9954307d-32c0-4c02-bf62-bbfe223b3c98 to /dev/sdb [ 1073.450182] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Preparing fetch location {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1073.450182] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Fetch image to [datastore2] OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583/OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583.vmdk {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1073.450366] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Downloading stream optimized image 173db284-6b8b-4355-bdde-6a9ba3be888e to [datastore2] OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583/OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583.vmdk on the data store datastore2 as vApp {{(pid=61985) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1073.450903] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Downloading image file data 173db284-6b8b-4355-bdde-6a9ba3be888e to the ESX as VM named 'OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583' {{(pid=61985) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1073.473992] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6a41f8-d76e-4966-b63d-be7d86f1d088 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.481104] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa310810-fe35-419e-8b80-8d1b708c0d9a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.511921] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.907s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.517881] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.260s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.519366] env[61985]: DEBUG nova.virt.block_device [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating existing volume attachment record: 72751b53-3e9f-43aa-beb8-95fbd0ca6a38 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1073.524929] env[61985]: INFO nova.network.neutron [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Port 0822504d-b73d-40b5-a7a2-4770f9d2f190 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1073.525398] env[61985]: DEBUG nova.network.neutron [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updating instance_info_cache with network_info: [{"id": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "address": "fa:16:3e:16:ec:4d", "network": {"id": "5014dbd4-285f-43e0-ad95-2107c34e808a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1627970140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2dbceebf07c43ac90ed80f8a09294b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88a63ff6-55", "ovs_interfaceid": "88a63ff6-55f6-4b8b-ba9e-188245f7ea87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.553539] env[61985]: INFO nova.scheduler.client.report [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted allocations for instance 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507 [ 1073.561413] env[61985]: DEBUG oslo_vmware.rw_handles [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1073.561413] env[61985]: value = "resgroup-9" [ 1073.561413] env[61985]: _type = "ResourcePool" [ 1073.561413] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1073.561413] env[61985]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-df32b1b2-d2ff-452b-bea1-acb473e7d278 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.587020] env[61985]: DEBUG oslo_vmware.rw_handles [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lease: (returnval){ [ 1073.587020] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c409c3-8475-c5f9-9b2c-a410281ece3c" [ 1073.587020] env[61985]: _type = "HttpNfcLease" [ 1073.587020] env[61985]: } obtained for vApp import into resource pool (val){ [ 1073.587020] env[61985]: value = "resgroup-9" [ 1073.587020] env[61985]: _type = "ResourcePool" [ 1073.587020] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1073.587020] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the lease: (returnval){ [ 1073.587020] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c409c3-8475-c5f9-9b2c-a410281ece3c" [ 1073.587020] env[61985]: _type = "HttpNfcLease" [ 1073.587020] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1073.595519] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1073.595519] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c409c3-8475-c5f9-9b2c-a410281ece3c" [ 1073.595519] env[61985]: _type = "HttpNfcLease" [ 1073.595519] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1073.705147] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936452, 'name': Rename_Task, 'duration_secs': 0.276542} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.705485] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1073.705772] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-304a3d78-50c6-46dc-acac-6dd743a39493 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.712661] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1073.712661] env[61985]: value = "task-936457" [ 1073.712661] env[61985]: _type = "Task" [ 1073.712661] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.718024] env[61985]: DEBUG nova.compute.manager [req-d7e264d7-9578-4f73-90b5-199cccb66e77 req-04b947f5-925b-46c3-9e62-71741f80b082 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Received event network-vif-plugged-1a081dc2-f497-4708-a3af-0b373d04cb85 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1073.718287] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7e264d7-9578-4f73-90b5-199cccb66e77 req-04b947f5-925b-46c3-9e62-71741f80b082 service nova] Acquiring lock "9079e605-e865-4eb1-9857-03769e4b5ba6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.718509] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7e264d7-9578-4f73-90b5-199cccb66e77 req-04b947f5-925b-46c3-9e62-71741f80b082 service nova] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.718745] env[61985]: DEBUG oslo_concurrency.lockutils [req-d7e264d7-9578-4f73-90b5-199cccb66e77 req-04b947f5-925b-46c3-9e62-71741f80b082 service nova] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.718933] env[61985]: DEBUG nova.compute.manager [req-d7e264d7-9578-4f73-90b5-199cccb66e77 req-04b947f5-925b-46c3-9e62-71741f80b082 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] No waiting events found dispatching network-vif-plugged-1a081dc2-f497-4708-a3af-0b373d04cb85 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1073.719119] env[61985]: WARNING nova.compute.manager [req-d7e264d7-9578-4f73-90b5-199cccb66e77 req-04b947f5-925b-46c3-9e62-71741f80b082 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Received unexpected event network-vif-plugged-1a081dc2-f497-4708-a3af-0b373d04cb85 for instance with vm_state building and task_state spawning. [ 1073.729219] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936457, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.767807] env[61985]: DEBUG nova.compute.manager [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1073.777322] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936450, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.882095] env[61985]: DEBUG nova.network.neutron [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Successfully updated port: 1a081dc2-f497-4708-a3af-0b373d04cb85 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1073.918908] env[61985]: DEBUG oslo_vmware.api [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936454, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.232019} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.919256] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1073.919427] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1073.919613] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1073.919813] env[61985]: INFO nova.compute.manager [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1073.920126] env[61985]: DEBUG oslo.service.loopingcall [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.921278] env[61985]: DEBUG nova.compute.manager [-] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1073.921278] env[61985]: DEBUG nova.network.neutron [-] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1074.031689] env[61985]: DEBUG oslo_concurrency.lockutils [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Releasing lock "refresh_cache-8db82979-e527-4c63-8ce8-fe1456f7708c" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.060728] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e0de8c04-e5b7-4cfd-9a4f-2322b7ab7a8e tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "65caa6d1-6e6e-4ab3-bb80-4c7e3b380507" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.177s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.102025] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1074.102025] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c409c3-8475-c5f9-9b2c-a410281ece3c" [ 1074.102025] env[61985]: _type = "HttpNfcLease" [ 1074.102025] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1074.225968] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936457, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.230445] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Didn't find any instances for network info cache update. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10018}} [ 1074.231040] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.231309] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.231540] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.231816] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.232103] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.232319] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.232526] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 1074.232760] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.277268] env[61985]: DEBUG oslo_vmware.api [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936450, 'name': PowerOnVM_Task, 'duration_secs': 1.240711} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.277551] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1074.277813] env[61985]: INFO nova.compute.manager [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Took 9.55 seconds to spawn the instance on the hypervisor. [ 1074.278020] env[61985]: DEBUG nova.compute.manager [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.280920] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d7a783-a95d-44cc-9eff-773925439c16 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.295449] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca76450d-7adc-4a6b-8095-2bd3e9384340 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.299151] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.303899] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458dc584-5ceb-4f9e-ba41-42ccd89b15d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.335215] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec7576a-f650-4666-a353-88272d3afa5c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.344109] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca955cba-cdb2-4550-a05e-6f5ade459c81 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.359260] env[61985]: DEBUG nova.compute.provider_tree [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.385414] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "refresh_cache-9079e605-e865-4eb1-9857-03769e4b5ba6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.385660] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "refresh_cache-9079e605-e865-4eb1-9857-03769e4b5ba6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.385807] env[61985]: DEBUG nova.network.neutron [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1074.536462] env[61985]: DEBUG oslo_concurrency.lockutils [None req-49e96dfc-cd1b-4a42-8c03-f6771505eed3 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "interface-8db82979-e527-4c63-8ce8-fe1456f7708c-0822504d-b73d-40b5-a7a2-4770f9d2f190" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.873s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.599031] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1074.599031] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c409c3-8475-c5f9-9b2c-a410281ece3c" [ 1074.599031] env[61985]: _type = "HttpNfcLease" [ 1074.599031] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1074.599279] env[61985]: DEBUG oslo_vmware.rw_handles [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1074.599279] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c409c3-8475-c5f9-9b2c-a410281ece3c" [ 1074.599279] env[61985]: _type = "HttpNfcLease" [ 1074.599279] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1074.600353] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e22b9a8-7efb-4da5-8913-dfd40f8f5924 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.608574] env[61985]: DEBUG oslo_vmware.rw_handles [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254393f-07b1-88f0-c2b8-c05e7d3bd4ff/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1074.608789] env[61985]: DEBUG oslo_vmware.rw_handles [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Creating HTTP connection to write to file with size = 31665152 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254393f-07b1-88f0-c2b8-c05e7d3bd4ff/disk-0.vmdk. {{(pid=61985) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1074.684014] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b240fe8b-8c48-4b65-8b1a-133ff7509268 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.723114] env[61985]: DEBUG oslo_vmware.api [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936457, 'name': PowerOnVM_Task, 'duration_secs': 0.555728} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.723397] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1074.723610] env[61985]: INFO nova.compute.manager [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Took 14.23 seconds to spawn the instance on the hypervisor. [ 1074.723799] env[61985]: DEBUG nova.compute.manager [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1074.724594] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11121062-48a7-4ce2-97fa-2f0b650803bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.737333] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.799837] env[61985]: INFO nova.compute.manager [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Took 25.16 seconds to build instance. [ 1074.861848] env[61985]: DEBUG nova.scheduler.client.report [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.922141] env[61985]: DEBUG nova.network.neutron [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1075.084133] env[61985]: DEBUG nova.network.neutron [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Updating instance_info_cache with network_info: [{"id": "1a081dc2-f497-4708-a3af-0b373d04cb85", "address": "fa:16:3e:5c:a1:b1", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a081dc2-f4", "ovs_interfaceid": "1a081dc2-f497-4708-a3af-0b373d04cb85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.248531] env[61985]: INFO nova.compute.manager [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Took 26.06 seconds to build instance. [ 1075.303878] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2fd88ed-a271-44c3-9ab0-74da1f234050 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "decfe088-90d6-49f9-9eae-7aa461219eaf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.679s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.345873] env[61985]: DEBUG nova.network.neutron [-] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.547384] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.547384] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.587315] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "refresh_cache-9079e605-e865-4eb1-9857-03769e4b5ba6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.587964] env[61985]: DEBUG nova.compute.manager [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Instance network_info: |[{"id": "1a081dc2-f497-4708-a3af-0b373d04cb85", "address": "fa:16:3e:5c:a1:b1", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a081dc2-f4", "ovs_interfaceid": "1a081dc2-f497-4708-a3af-0b373d04cb85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1075.588660] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:a1:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd829efb7-e98e-4b67-bd03-b0888287dbfd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a081dc2-f497-4708-a3af-0b373d04cb85', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1075.601737] env[61985]: DEBUG oslo.service.loopingcall [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.602627] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1075.603022] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c4d7b5a-e201-443d-95aa-4857c1ca0d64 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.634519] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1075.634519] env[61985]: value = "task-936460" [ 1075.634519] env[61985]: _type = "Task" [ 1075.634519] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.650773] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936460, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.753991] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d6f95e3e-a37d-4f5a-b07f-dc448c16d25a tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.576s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.757972] env[61985]: DEBUG nova.compute.manager [req-e0d14eec-d8a1-4665-badd-ca3eed78ac33 req-2b62711f-f4ec-4663-b8d0-d91040e9a635 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Received event network-changed-1a081dc2-f497-4708-a3af-0b373d04cb85 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1075.758582] env[61985]: DEBUG nova.compute.manager [req-e0d14eec-d8a1-4665-badd-ca3eed78ac33 req-2b62711f-f4ec-4663-b8d0-d91040e9a635 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Refreshing instance network info cache due to event network-changed-1a081dc2-f497-4708-a3af-0b373d04cb85. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1075.758790] env[61985]: DEBUG oslo_concurrency.lockutils [req-e0d14eec-d8a1-4665-badd-ca3eed78ac33 req-2b62711f-f4ec-4663-b8d0-d91040e9a635 service nova] Acquiring lock "refresh_cache-9079e605-e865-4eb1-9857-03769e4b5ba6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.759018] env[61985]: DEBUG oslo_concurrency.lockutils [req-e0d14eec-d8a1-4665-badd-ca3eed78ac33 req-2b62711f-f4ec-4663-b8d0-d91040e9a635 service nova] Acquired lock "refresh_cache-9079e605-e865-4eb1-9857-03769e4b5ba6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.759202] env[61985]: DEBUG nova.network.neutron [req-e0d14eec-d8a1-4665-badd-ca3eed78ac33 req-2b62711f-f4ec-4663-b8d0-d91040e9a635 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Refreshing network info cache for port 1a081dc2-f497-4708-a3af-0b373d04cb85 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1075.805825] env[61985]: DEBUG oslo_concurrency.lockutils [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "8eea60eb-93cf-4701-9ba3-0805691ca016" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.806268] env[61985]: DEBUG oslo_concurrency.lockutils [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.806584] env[61985]: DEBUG oslo_concurrency.lockutils [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.806859] env[61985]: DEBUG oslo_concurrency.lockutils [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.807225] env[61985]: DEBUG oslo_concurrency.lockutils [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.813172] env[61985]: INFO nova.compute.manager [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Terminating instance [ 1075.821767] env[61985]: DEBUG nova.compute.manager [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1075.821926] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1075.822902] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0abb6816-8312-4717-b19a-d66337b5ee62 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.833052] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1075.833393] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d55fe541-f4be-4ef2-8581-e6c15ab0387f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.840319] env[61985]: DEBUG oslo_vmware.api [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1075.840319] env[61985]: value = "task-936461" [ 1075.840319] env[61985]: _type = "Task" [ 1075.840319] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.848404] env[61985]: INFO nova.compute.manager [-] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Took 1.93 seconds to deallocate network for instance. [ 1075.857542] env[61985]: DEBUG oslo_vmware.api [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936461, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.873081] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.355s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.879521] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.580s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.881600] env[61985]: INFO nova.compute.claims [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.052252] env[61985]: DEBUG nova.compute.manager [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1076.152809] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936460, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.176655] env[61985]: DEBUG oslo_vmware.rw_handles [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Completed reading data from the image iterator. {{(pid=61985) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1076.176897] env[61985]: DEBUG oslo_vmware.rw_handles [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254393f-07b1-88f0-c2b8-c05e7d3bd4ff/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1076.177951] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d38f9677-3fc9-4d87-84b1-582c711e02ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.187141] env[61985]: DEBUG oslo_vmware.rw_handles [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254393f-07b1-88f0-c2b8-c05e7d3bd4ff/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1076.187141] env[61985]: DEBUG oslo_vmware.rw_handles [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254393f-07b1-88f0-c2b8-c05e7d3bd4ff/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1076.187141] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-0b7898b6-476a-409f-a311-6df4b6640fb5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.350504] env[61985]: DEBUG oslo_vmware.api [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936461, 'name': PowerOffVM_Task, 'duration_secs': 0.393321} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.350870] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1076.351211] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1076.351333] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d01c4474-48d9-4893-8c8d-85dc7dbab353 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.368698] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.475466] env[61985]: INFO nova.scheduler.client.report [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted allocation for migration 3e248959-8537-41af-b11e-32f60bab4249 [ 1076.532657] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1076.532932] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1076.533185] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Deleting the datastore file [datastore1] 8eea60eb-93cf-4701-9ba3-0805691ca016 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1076.533505] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79262a26-25d4-41ba-aab8-6097c8d25d71 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.542369] env[61985]: DEBUG oslo_vmware.api [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1076.542369] env[61985]: value = "task-936464" [ 1076.542369] env[61985]: _type = "Task" [ 1076.542369] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.553591] env[61985]: DEBUG oslo_vmware.api [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936464, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.559501] env[61985]: DEBUG oslo_vmware.rw_handles [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5254393f-07b1-88f0-c2b8-c05e7d3bd4ff/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1076.559729] env[61985]: INFO nova.virt.vmwareapi.images [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Downloaded image file data 173db284-6b8b-4355-bdde-6a9ba3be888e [ 1076.562715] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-408522e3-6ea0-4d62-a8aa-52a6a5d76f6c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.579633] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-84d671d5-69f5-44bb-a45c-76b0f1739fb2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.588049] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.649010] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936460, 'name': CreateVM_Task, 'duration_secs': 0.612139} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.650278] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1076.650598] env[61985]: INFO nova.virt.vmwareapi.images [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] The imported VM was unregistered [ 1076.653391] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Caching image {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1076.653632] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Creating directory with path [datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1076.654361] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.654530] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.654877] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1076.655149] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-198c8d2b-25dc-498d-bed2-5f3c51403af4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.657371] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b95a1b82-6630-4c23-a59e-7a6d1d6a074e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.663156] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1076.663156] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5201f778-d4da-041a-f6ec-d2eae000afc0" [ 1076.663156] env[61985]: _type = "Task" [ 1076.663156] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.674908] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5201f778-d4da-041a-f6ec-d2eae000afc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.676281] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Created directory with path [datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1076.676710] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583/OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583.vmdk to [datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e/173db284-6b8b-4355-bdde-6a9ba3be888e.vmdk. {{(pid=61985) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1076.677510] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d1ea82e2-412c-4cff-b6df-b5a6731ee6e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.684730] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1076.684730] env[61985]: value = "task-936466" [ 1076.684730] env[61985]: _type = "Task" [ 1076.684730] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.698992] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936466, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.730014] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "decfe088-90d6-49f9-9eae-7aa461219eaf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.731251] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "decfe088-90d6-49f9-9eae-7aa461219eaf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.731842] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "decfe088-90d6-49f9-9eae-7aa461219eaf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.732116] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "decfe088-90d6-49f9-9eae-7aa461219eaf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.732321] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "decfe088-90d6-49f9-9eae-7aa461219eaf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.738567] env[61985]: INFO nova.compute.manager [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Terminating instance [ 1076.743415] env[61985]: DEBUG nova.compute.manager [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1076.743972] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1076.745191] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdebe1a-ec1d-427e-862e-67372ca31885 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.756914] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1076.757402] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b02f2fd-9ed8-4ce5-a448-45d33b4f5b50 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.765867] env[61985]: DEBUG oslo_vmware.api [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1076.765867] env[61985]: value = "task-936467" [ 1076.765867] env[61985]: _type = "Task" [ 1076.765867] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.776401] env[61985]: DEBUG oslo_vmware.api [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.902854] env[61985]: DEBUG nova.network.neutron [req-e0d14eec-d8a1-4665-badd-ca3eed78ac33 req-2b62711f-f4ec-4663-b8d0-d91040e9a635 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Updated VIF entry in instance network info cache for port 1a081dc2-f497-4708-a3af-0b373d04cb85. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1076.903444] env[61985]: DEBUG nova.network.neutron [req-e0d14eec-d8a1-4665-badd-ca3eed78ac33 req-2b62711f-f4ec-4663-b8d0-d91040e9a635 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Updating instance_info_cache with network_info: [{"id": "1a081dc2-f497-4708-a3af-0b373d04cb85", "address": "fa:16:3e:5c:a1:b1", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a081dc2-f4", "ovs_interfaceid": "1a081dc2-f497-4708-a3af-0b373d04cb85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.987429] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d964c398-eb89-47da-9fbb-1d7bf56bb094 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.830s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.057043] env[61985]: DEBUG oslo_vmware.api [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936464, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.373197} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.058236] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1077.058569] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1077.059035] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1077.059331] env[61985]: INFO nova.compute.manager [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Took 1.24 seconds to destroy the instance on the hypervisor. [ 1077.059669] env[61985]: DEBUG oslo.service.loopingcall [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1077.059980] env[61985]: DEBUG nova.compute.manager [-] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1077.060144] env[61985]: DEBUG nova.network.neutron [-] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1077.183222] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5201f778-d4da-041a-f6ec-d2eae000afc0, 'name': SearchDatastore_Task, 'duration_secs': 0.024354} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.183604] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.184094] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1077.184205] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.184331] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.184580] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.184877] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2b073ba-8b1e-4261-9b42-1f2c22b8d93d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.201759] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936466, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.203024] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdbaa24e-4bc7-48cc-bfba-67c0a31eee01 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.208433] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.208647] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1077.213118] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd12c20e-ee94-419e-8106-b83a72040ae5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.217251] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7fb7c6-f2eb-4cf4-a2d2-7f42e804ca11 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.225139] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1077.225139] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529e67a8-3356-1d70-d7ad-a2a863c287b4" [ 1077.225139] env[61985]: _type = "Task" [ 1077.225139] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.266435] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65095f8-aea9-4e35-926b-bbb29127a158 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.280553] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529e67a8-3356-1d70-d7ad-a2a863c287b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.286257] env[61985]: DEBUG oslo_vmware.api [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936467, 'name': PowerOffVM_Task, 'duration_secs': 0.234851} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.287687] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb69b160-75c9-4ce3-aca6-472d232f2f2d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.292122] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1077.292295] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1077.292603] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fcc1968b-7e6f-40ff-be14-8a2d40d73e9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.306771] env[61985]: DEBUG nova.compute.provider_tree [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.361129] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1077.361502] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1077.361656] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleting the datastore file [datastore1] decfe088-90d6-49f9-9eae-7aa461219eaf {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1077.361935] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fabf3a0d-7667-4b5e-b54a-810440da5198 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.369533] env[61985]: DEBUG oslo_vmware.api [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1077.369533] env[61985]: value = "task-936469" [ 1077.369533] env[61985]: _type = "Task" [ 1077.369533] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.379116] env[61985]: DEBUG oslo_vmware.api [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936469, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.406514] env[61985]: DEBUG oslo_concurrency.lockutils [req-e0d14eec-d8a1-4665-badd-ca3eed78ac33 req-2b62711f-f4ec-4663-b8d0-d91040e9a635 service nova] Releasing lock "refresh_cache-9079e605-e865-4eb1-9857-03769e4b5ba6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.406842] env[61985]: DEBUG nova.compute.manager [req-e0d14eec-d8a1-4665-badd-ca3eed78ac33 req-2b62711f-f4ec-4663-b8d0-d91040e9a635 service nova] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Received event network-vif-deleted-88a63ff6-55f6-4b8b-ba9e-188245f7ea87 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1077.698393] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936466, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.775133] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529e67a8-3356-1d70-d7ad-a2a863c287b4, 'name': SearchDatastore_Task, 'duration_secs': 0.082641} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.775818] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5459f979-b451-4948-b088-ad50f5fb05f0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.783047] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1077.783047] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52877442-15e8-eef6-9954-861575cb265c" [ 1077.783047] env[61985]: _type = "Task" [ 1077.783047] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.791838] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52877442-15e8-eef6-9954-861575cb265c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.799462] env[61985]: DEBUG nova.compute.manager [req-6467a79f-0421-48a8-a9b1-8da57037575d req-db485b66-abab-48d1-b702-c7fe4bcbd556 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received event network-vif-deleted-c4f26df9-9705-48e3-8414-595fed68fdee {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1077.799687] env[61985]: INFO nova.compute.manager [req-6467a79f-0421-48a8-a9b1-8da57037575d req-db485b66-abab-48d1-b702-c7fe4bcbd556 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Neutron deleted interface c4f26df9-9705-48e3-8414-595fed68fdee; detaching it from the instance and deleting it from the info cache [ 1077.800103] env[61985]: DEBUG nova.network.neutron [req-6467a79f-0421-48a8-a9b1-8da57037575d req-db485b66-abab-48d1-b702-c7fe4bcbd556 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Updating instance_info_cache with network_info: [{"id": "01e88a1a-79a9-4c6f-a939-607ae4ecd7f2", "address": "fa:16:3e:08:9d:22", "network": {"id": "94f9d52d-2ccd-44f4-85f0-c6c953a38dec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1611261022", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.202", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01e88a1a-79", "ovs_interfaceid": "01e88a1a-79a9-4c6f-a939-607ae4ecd7f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a2e1f2a-6455-4354-83df-19ac0b8f17db", "address": "fa:16:3e:78:f7:aa", "network": {"id": "94f9d52d-2ccd-44f4-85f0-c6c953a38dec", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1611261022", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.174", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a2e1f2a-64", "ovs_interfaceid": "6a2e1f2a-6455-4354-83df-19ac0b8f17db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.809426] env[61985]: DEBUG nova.scheduler.client.report [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1077.885283] env[61985]: DEBUG oslo_vmware.api [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936469, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.23075} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.886331] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1077.886628] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1077.886836] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1077.887264] env[61985]: INFO nova.compute.manager [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1077.887589] env[61985]: DEBUG oslo.service.loopingcall [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1077.887848] env[61985]: DEBUG nova.compute.manager [-] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1077.887949] env[61985]: DEBUG nova.network.neutron [-] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1078.070226] env[61985]: DEBUG oslo_concurrency.lockutils [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "8c4a1fe4-4ded-4064-8261-24826181a68c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.070226] env[61985]: DEBUG oslo_concurrency.lockutils [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.070226] env[61985]: DEBUG oslo_concurrency.lockutils [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "8c4a1fe4-4ded-4064-8261-24826181a68c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.070226] env[61985]: DEBUG oslo_concurrency.lockutils [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.070226] env[61985]: DEBUG oslo_concurrency.lockutils [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.073130] env[61985]: INFO nova.compute.manager [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Terminating instance [ 1078.078329] env[61985]: DEBUG nova.compute.manager [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1078.078329] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1078.082523] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8382faa-08da-4b25-a661-3d6d2c79f214 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.091737] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1078.092154] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b9bc220-c2c1-4590-8900-61afd4c5b8e8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.099146] env[61985]: DEBUG oslo_vmware.api [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1078.099146] env[61985]: value = "task-936470" [ 1078.099146] env[61985]: _type = "Task" [ 1078.099146] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.107798] env[61985]: DEBUG oslo_vmware.api [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936470, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.201714] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936466, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.297418] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52877442-15e8-eef6-9954-861575cb265c, 'name': SearchDatastore_Task, 'duration_secs': 0.085346} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.299687] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.299903] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 9079e605-e865-4eb1-9857-03769e4b5ba6/9079e605-e865-4eb1-9857-03769e4b5ba6.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1078.304189] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-463ba840-07d5-4e77-9185-f97af26e6f3b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.307948] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f2263f5-169c-492f-a27f-4a83fe7c4213 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.319767] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030399ed-cb40-42e4-9bc8-1ca026347bfb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.332274] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.332606] env[61985]: DEBUG nova.compute.manager [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1078.335343] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1078.335343] env[61985]: value = "task-936471" [ 1078.335343] env[61985]: _type = "Task" [ 1078.335343] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.336250] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.599s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.336250] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.337102] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1078.337102] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.969s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.337102] env[61985]: DEBUG nova.objects.instance [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'resources' on Instance uuid 8db82979-e527-4c63-8ce8-fe1456f7708c {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.339710] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753fb1a5-793c-432c-9f0f-9c2beb5fdaf1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.363482] env[61985]: DEBUG nova.compute.manager [req-6467a79f-0421-48a8-a9b1-8da57037575d req-db485b66-abab-48d1-b702-c7fe4bcbd556 service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Detach interface failed, port_id=c4f26df9-9705-48e3-8414-595fed68fdee, reason: Instance 8eea60eb-93cf-4701-9ba3-0805691ca016 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1078.369144] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae53e8e-f45a-415c-a1d9-ca8a509bff20 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.373873] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936471, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.377490] env[61985]: DEBUG nova.compute.manager [req-5d4e2ddb-47a9-4e32-87b1-b1b3b049a341 req-8a94490c-3d4f-4bc2-8343-c14ed016f235 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Received event network-vif-deleted-5a9c3e06-ff99-444a-a9ed-8f508e1854fb {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1078.377869] env[61985]: INFO nova.compute.manager [req-5d4e2ddb-47a9-4e32-87b1-b1b3b049a341 req-8a94490c-3d4f-4bc2-8343-c14ed016f235 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Neutron deleted interface 5a9c3e06-ff99-444a-a9ed-8f508e1854fb; detaching it from the instance and deleting it from the info cache [ 1078.378193] env[61985]: DEBUG nova.network.neutron [req-5d4e2ddb-47a9-4e32-87b1-b1b3b049a341 req-8a94490c-3d4f-4bc2-8343-c14ed016f235 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.393700] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b094d33d-b803-4a2c-bd5d-b7df170933c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.416145] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea12117c-92ad-491c-8f0e-199123b5aa6f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.453026] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180175MB free_disk=43GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1078.453026] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.595300] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1078.595597] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211465', 'volume_id': '9954307d-32c0-4c02-bf62-bbfe223b3c98', 'name': 'volume-9954307d-32c0-4c02-bf62-bbfe223b3c98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '885a36d9-24c6-407a-9d6b-01cd97e90674', 'attached_at': '', 'detached_at': '', 'volume_id': '9954307d-32c0-4c02-bf62-bbfe223b3c98', 'serial': '9954307d-32c0-4c02-bf62-bbfe223b3c98'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1078.596520] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1c8460-f1f3-4f5c-ab38-2d7146e9f86b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.609302] env[61985]: DEBUG oslo_vmware.api [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936470, 'name': PowerOffVM_Task, 'duration_secs': 0.268347} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.622582] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1078.622814] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1078.624841] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e79827d-4292-46a0-885a-db1c4914b6a0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.627980] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908b49b2-8c1f-44f3-9dd3-4c1cc52bf9bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.662252] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] volume-9954307d-32c0-4c02-bf62-bbfe223b3c98/volume-9954307d-32c0-4c02-bf62-bbfe223b3c98.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.667097] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c83680d6-f9e8-48ee-8f0f-9d993a739e75 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.687206] env[61985]: DEBUG oslo_vmware.api [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1078.687206] env[61985]: value = "task-936473" [ 1078.687206] env[61985]: _type = "Task" [ 1078.687206] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.703725] env[61985]: DEBUG oslo_vmware.api [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936473, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.707888] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936466, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.759103] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1078.759103] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1078.759103] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleting the datastore file [datastore1] 8c4a1fe4-4ded-4064-8261-24826181a68c {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1078.759103] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0246d29a-b7f9-436a-a836-707296baf519 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.768047] env[61985]: DEBUG oslo_vmware.api [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1078.768047] env[61985]: value = "task-936474" [ 1078.768047] env[61985]: _type = "Task" [ 1078.768047] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.772873] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d9a749-315d-44c9-8f50-0b05f71cc7c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.781463] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcac6d79-62b3-4cb6-a8f8-76d03d7bedc4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.788658] env[61985]: DEBUG oslo_vmware.api [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936474, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.816931] env[61985]: DEBUG nova.network.neutron [-] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.818934] env[61985]: DEBUG nova.network.neutron [-] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.821396] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05dda1bb-ebab-4a51-8917-f00b305ac158 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.829879] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e13d681-6aca-48f4-b37e-01e2d2dd8d36 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.841399] env[61985]: DEBUG nova.compute.utils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1078.842696] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "926a2b59-8614-4e71-9d94-fd9feefdef46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.843158] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "926a2b59-8614-4e71-9d94-fd9feefdef46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.856448] env[61985]: DEBUG nova.compute.manager [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1078.856635] env[61985]: DEBUG nova.network.neutron [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1078.861408] env[61985]: DEBUG nova.compute.provider_tree [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.867685] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936471, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.882065] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-25c568fc-e278-476e-a569-fe2cd6461a88 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.893651] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b2bb74-e393-43c8-9f5d-5edf03a18ac0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.915821] env[61985]: DEBUG nova.policy [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fdf7c32304a6427daa1d611353f11b4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2504537f92814207ab6453b6aea4f1fb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1078.930283] env[61985]: DEBUG nova.compute.manager [req-5d4e2ddb-47a9-4e32-87b1-b1b3b049a341 req-8a94490c-3d4f-4bc2-8343-c14ed016f235 service nova] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Detach interface failed, port_id=5a9c3e06-ff99-444a-a9ed-8f508e1854fb, reason: Instance decfe088-90d6-49f9-9eae-7aa461219eaf could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1079.209120] env[61985]: DEBUG oslo_vmware.api [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936473, 'name': ReconfigVM_Task, 'duration_secs': 0.481097} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.213849] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfigured VM instance instance-0000005a to attach disk [datastore2] volume-9954307d-32c0-4c02-bf62-bbfe223b3c98/volume-9954307d-32c0-4c02-bf62-bbfe223b3c98.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1079.222838] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936466, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.223239] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc2f6244-f6c3-4952-8221-d78e6b51fbf8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.249709] env[61985]: DEBUG oslo_vmware.api [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1079.249709] env[61985]: value = "task-936475" [ 1079.249709] env[61985]: _type = "Task" [ 1079.249709] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.263128] env[61985]: DEBUG oslo_vmware.api [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936475, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.281619] env[61985]: DEBUG oslo_vmware.api [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936474, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27602} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.281834] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.282955] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1079.283195] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1079.283393] env[61985]: INFO nova.compute.manager [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1079.284367] env[61985]: DEBUG oslo.service.loopingcall [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.284367] env[61985]: DEBUG nova.compute.manager [-] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1079.284367] env[61985]: DEBUG nova.network.neutron [-] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1079.320286] env[61985]: INFO nova.compute.manager [-] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Took 1.43 seconds to deallocate network for instance. [ 1079.320867] env[61985]: INFO nova.compute.manager [-] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Took 2.26 seconds to deallocate network for instance. [ 1079.345402] env[61985]: DEBUG nova.compute.manager [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1079.353263] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936471, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.360569] env[61985]: DEBUG nova.network.neutron [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Successfully created port: 88de6684-404d-48bc-85c9-214ab6d77e42 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1079.362754] env[61985]: DEBUG nova.compute.manager [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1079.366303] env[61985]: DEBUG nova.scheduler.client.report [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.706893] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936466, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.762905] env[61985]: DEBUG oslo_vmware.api [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936475, 'name': ReconfigVM_Task, 'duration_secs': 0.143054} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.763553] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211465', 'volume_id': '9954307d-32c0-4c02-bf62-bbfe223b3c98', 'name': 'volume-9954307d-32c0-4c02-bf62-bbfe223b3c98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '885a36d9-24c6-407a-9d6b-01cd97e90674', 'attached_at': '', 'detached_at': '', 'volume_id': '9954307d-32c0-4c02-bf62-bbfe223b3c98', 'serial': '9954307d-32c0-4c02-bf62-bbfe223b3c98'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1079.832416] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.833213] env[61985]: DEBUG oslo_concurrency.lockutils [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.840301] env[61985]: DEBUG nova.compute.manager [req-8ee1d7b4-c14b-4ed4-aaf3-c1c5fab9c5e3 req-91682b08-c8ca-427f-9efc-dcb752f64b4b service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received event network-vif-deleted-01e88a1a-79a9-4c6f-a939-607ae4ecd7f2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1079.841083] env[61985]: DEBUG nova.compute.manager [req-8ee1d7b4-c14b-4ed4-aaf3-c1c5fab9c5e3 req-91682b08-c8ca-427f-9efc-dcb752f64b4b service nova] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Received event network-vif-deleted-6a2e1f2a-6455-4354-83df-19ac0b8f17db {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1079.856291] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936471, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.875052] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.538s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.883347] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.292s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.883347] env[61985]: INFO nova.compute.claims [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1079.909608] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.912895] env[61985]: INFO nova.scheduler.client.report [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Deleted allocations for instance 8db82979-e527-4c63-8ce8-fe1456f7708c [ 1080.145274] env[61985]: DEBUG nova.network.neutron [-] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.206044] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936466, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.084495} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.206291] env[61985]: INFO nova.virt.vmwareapi.ds_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583/OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583.vmdk to [datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e/173db284-6b8b-4355-bdde-6a9ba3be888e.vmdk. [ 1080.206499] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Cleaning up location [datastore2] OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1080.206737] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_9cbe87c7-be68-4839-b6fc-174e5f611583 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.207010] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35b89bb7-57c3-45b7-b89d-0fb1b57b00a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.212836] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1080.212836] env[61985]: value = "task-936476" [ 1080.212836] env[61985]: _type = "Task" [ 1080.212836] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.220923] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936476, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.352634] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936471, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.847125} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.352948] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 9079e605-e865-4eb1-9857-03769e4b5ba6/9079e605-e865-4eb1-9857-03769e4b5ba6.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1080.353206] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1080.353465] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3e2302d-5564-421a-843d-843b01502992 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.360821] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1080.360821] env[61985]: value = "task-936477" [ 1080.360821] env[61985]: _type = "Task" [ 1080.360821] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.361940] env[61985]: DEBUG nova.compute.manager [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1080.372649] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936477, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.384100] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.384287] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.384422] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.384533] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.384682] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.384832] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.385061] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.385237] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.385410] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.385575] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.385763] env[61985]: DEBUG nova.virt.hardware [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.386878] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1177167c-b4cd-4da9-8976-313a21a1978a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.396903] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee720b9-0956-46e9-b113-fd5ce3b5da89 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.404396] env[61985]: DEBUG nova.compute.manager [req-4775894a-1378-46bd-acdc-9cc6643d12d6 req-1769b7c8-d1f0-460e-80cc-1a7cc5f94ed0 service nova] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Received event network-vif-deleted-88467e26-c1c7-4b6a-ab2b-9d86b552c0c1 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1080.422708] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ac71233c-35c6-4ca2-a209-8f596df6dd5c tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "8db82979-e527-4c63-8ce8-fe1456f7708c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.689s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.650530] env[61985]: INFO nova.compute.manager [-] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Took 1.37 seconds to deallocate network for instance. [ 1080.724837] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936476, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.040573} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.725134] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1080.725308] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "[datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e/173db284-6b8b-4355-bdde-6a9ba3be888e.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.725572] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e/173db284-6b8b-4355-bdde-6a9ba3be888e.vmdk to [datastore2] 87006fbf-ea90-4d9a-88af-001de424ac14/87006fbf-ea90-4d9a-88af-001de424ac14.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1080.725833] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4f7eb87-a633-4fba-a121-528a91e780b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.732287] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1080.732287] env[61985]: value = "task-936478" [ 1080.732287] env[61985]: _type = "Task" [ 1080.732287] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.741104] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936478, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.809135] env[61985]: DEBUG nova.objects.instance [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'flavor' on Instance uuid 885a36d9-24c6-407a-9d6b-01cd97e90674 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.834188] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.834435] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.871123] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936477, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082843} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.871440] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1080.872248] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd6a05c-30b0-40f2-bfb1-f7dd8333a00d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.895364] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 9079e605-e865-4eb1-9857-03769e4b5ba6/9079e605-e865-4eb1-9857-03769e4b5ba6.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1080.898579] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f96e0b95-2a0a-4e02-b358-ccd8b48e07fd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.920451] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1080.920451] env[61985]: value = "task-936479" [ 1080.920451] env[61985]: _type = "Task" [ 1080.920451] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.929653] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936479, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.028054] env[61985]: DEBUG nova.network.neutron [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Successfully updated port: 88de6684-404d-48bc-85c9-214ab6d77e42 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1081.133483] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39fe7c4-fc6c-4609-9782-3103e23902c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.141215] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052448fc-e40d-4da7-b094-a8de2ae52177 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.174241] env[61985]: DEBUG oslo_concurrency.lockutils [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.175641] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91f80f1-20e2-404b-9525-819c7891fd00 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.178404] env[61985]: DEBUG oslo_concurrency.lockutils [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.178633] env[61985]: DEBUG oslo_concurrency.lockutils [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.178840] env[61985]: DEBUG oslo_concurrency.lockutils [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.179054] env[61985]: DEBUG oslo_concurrency.lockutils [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.179248] env[61985]: DEBUG oslo_concurrency.lockutils [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.181501] env[61985]: INFO nova.compute.manager [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Terminating instance [ 1081.185706] env[61985]: DEBUG nova.compute.manager [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1081.185706] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1081.186761] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030c714c-4838-4ba1-9121-eaaafbb0d583 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.191592] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea54cb3-1aab-4d44-8970-ce831b158aea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.199053] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1081.206632] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17be7b06-1634-4ed0-87a6-743414b24499 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.208751] env[61985]: DEBUG nova.compute.provider_tree [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.214127] env[61985]: DEBUG oslo_vmware.api [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1081.214127] env[61985]: value = "task-936480" [ 1081.214127] env[61985]: _type = "Task" [ 1081.214127] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.225559] env[61985]: DEBUG oslo_vmware.api [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.242349] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936478, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.313861] env[61985]: DEBUG oslo_concurrency.lockutils [None req-31170f80-3ed8-4cad-b7f0-c8bad27ef1f3 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.884s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.337651] env[61985]: DEBUG nova.compute.manager [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1081.432787] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936479, 'name': ReconfigVM_Task, 'duration_secs': 0.290242} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.433116] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 9079e605-e865-4eb1-9857-03769e4b5ba6/9079e605-e865-4eb1-9857-03769e4b5ba6.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1081.433845] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10a6d4a4-3d23-42a8-a7b3-f75d216928a9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.441968] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1081.441968] env[61985]: value = "task-936481" [ 1081.441968] env[61985]: _type = "Task" [ 1081.441968] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.451365] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936481, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.531274] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquiring lock "refresh_cache-acebe0a7-62d4-48d4-a1f7-5df0fa465753" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.531567] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquired lock "refresh_cache-acebe0a7-62d4-48d4-a1f7-5df0fa465753" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.531851] env[61985]: DEBUG nova.network.neutron [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1081.713057] env[61985]: DEBUG nova.scheduler.client.report [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.729219] env[61985]: DEBUG oslo_vmware.api [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.746570] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936478, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.862761] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.871378] env[61985]: DEBUG nova.compute.manager [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Received event network-vif-plugged-88de6684-404d-48bc-85c9-214ab6d77e42 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1081.871598] env[61985]: DEBUG oslo_concurrency.lockutils [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] Acquiring lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.871813] env[61985]: DEBUG oslo_concurrency.lockutils [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] Lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.871995] env[61985]: DEBUG oslo_concurrency.lockutils [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] Lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.872185] env[61985]: DEBUG nova.compute.manager [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] No waiting events found dispatching network-vif-plugged-88de6684-404d-48bc-85c9-214ab6d77e42 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1081.872357] env[61985]: WARNING nova.compute.manager [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Received unexpected event network-vif-plugged-88de6684-404d-48bc-85c9-214ab6d77e42 for instance with vm_state building and task_state spawning. [ 1081.872522] env[61985]: DEBUG nova.compute.manager [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Received event network-changed-88de6684-404d-48bc-85c9-214ab6d77e42 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1081.872679] env[61985]: DEBUG nova.compute.manager [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Refreshing instance network info cache due to event network-changed-88de6684-404d-48bc-85c9-214ab6d77e42. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1081.872848] env[61985]: DEBUG oslo_concurrency.lockutils [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] Acquiring lock "refresh_cache-acebe0a7-62d4-48d4-a1f7-5df0fa465753" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.953520] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936481, 'name': Rename_Task, 'duration_secs': 0.162556} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.953809] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1081.954086] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-983c1daa-a343-4952-9b63-b6ff601bc3d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.961634] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1081.961634] env[61985]: value = "task-936482" [ 1081.961634] env[61985]: _type = "Task" [ 1081.961634] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.969934] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936482, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.069542] env[61985]: DEBUG nova.network.neutron [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1082.219143] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.219735] env[61985]: DEBUG nova.compute.manager [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1082.226765] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.775s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.244058] env[61985]: DEBUG nova.network.neutron [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Updating instance_info_cache with network_info: [{"id": "88de6684-404d-48bc-85c9-214ab6d77e42", "address": "fa:16:3e:93:90:1b", "network": {"id": "3e64e59c-d2d2-4d75-87a7-e63c5857c6a5", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1971303841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2504537f92814207ab6453b6aea4f1fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88de6684-40", "ovs_interfaceid": "88de6684-404d-48bc-85c9-214ab6d77e42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.245205] env[61985]: DEBUG oslo_vmware.api [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936480, 'name': PowerOffVM_Task, 'duration_secs': 0.879228} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.246330] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1082.246740] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1082.246841] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63cce70b-2339-47ab-8908-db6178ff4d6e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.252218] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936478, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.376450] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1082.376696] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1082.376888] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Deleting the datastore file [datastore2] 5f2aa808-8d1e-471e-9ef7-0f91590ec546 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1082.377219] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27a9f242-9e09-4757-a8da-72f23a66fee4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.388521] env[61985]: DEBUG oslo_vmware.api [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for the task: (returnval){ [ 1082.388521] env[61985]: value = "task-936484" [ 1082.388521] env[61985]: _type = "Task" [ 1082.388521] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.399604] env[61985]: DEBUG oslo_vmware.api [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936484, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.475658] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936482, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.683047] env[61985]: DEBUG nova.compute.manager [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Stashing vm_state: active {{(pid=61985) _prep_resize /opt/stack/nova/nova/compute/manager.py:5630}} [ 1082.738754] env[61985]: DEBUG nova.compute.utils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1082.740387] env[61985]: DEBUG nova.compute.manager [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1082.740559] env[61985]: DEBUG nova.network.neutron [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1082.746513] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Releasing lock "refresh_cache-acebe0a7-62d4-48d4-a1f7-5df0fa465753" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.746846] env[61985]: DEBUG nova.compute.manager [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Instance network_info: |[{"id": "88de6684-404d-48bc-85c9-214ab6d77e42", "address": "fa:16:3e:93:90:1b", "network": {"id": "3e64e59c-d2d2-4d75-87a7-e63c5857c6a5", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1971303841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2504537f92814207ab6453b6aea4f1fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88de6684-40", "ovs_interfaceid": "88de6684-404d-48bc-85c9-214ab6d77e42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1082.747697] env[61985]: DEBUG oslo_concurrency.lockutils [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] Acquired lock "refresh_cache-acebe0a7-62d4-48d4-a1f7-5df0fa465753" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.747858] env[61985]: DEBUG nova.network.neutron [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Refreshing network info cache for port 88de6684-404d-48bc-85c9-214ab6d77e42 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1082.749134] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:90:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88de6684-404d-48bc-85c9-214ab6d77e42', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1082.757380] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Creating folder: Project (2504537f92814207ab6453b6aea4f1fb). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1082.762241] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-70b4fa48-5e15-4667-b8b0-8890076c6bfc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.769038] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936478, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.777853] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Created folder: Project (2504537f92814207ab6453b6aea4f1fb) in parent group-v211285. [ 1082.778113] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Creating folder: Instances. Parent ref: group-v211467. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1082.778397] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf808308-397a-4292-a0ff-fba06c51ed46 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.787882] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Created folder: Instances in parent group-v211467. [ 1082.788198] env[61985]: DEBUG oslo.service.loopingcall [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1082.788411] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1082.788632] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a677a660-f772-4548-8fc3-ab07291b4c88 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.804630] env[61985]: DEBUG nova.policy [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faa19f385cd0426d8565e31e286800d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5625013599b44418bd56eb604e14be58', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1082.813581] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1082.813581] env[61985]: value = "task-936487" [ 1082.813581] env[61985]: _type = "Task" [ 1082.813581] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.822343] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936487, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.900334] env[61985]: DEBUG oslo_vmware.api [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936484, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.976221] env[61985]: DEBUG oslo_vmware.api [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936482, 'name': PowerOnVM_Task, 'duration_secs': 0.641864} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.976532] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1082.976678] env[61985]: INFO nova.compute.manager [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Took 10.61 seconds to spawn the instance on the hypervisor. [ 1082.976866] env[61985]: DEBUG nova.compute.manager [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1082.977853] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe9e4ce-879b-44fb-ad0c-f0879e78cd1c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.111217] env[61985]: DEBUG nova.network.neutron [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Successfully created port: e1f6f60d-06ee-469d-bf35-a3af36741d55 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1083.118442] env[61985]: DEBUG nova.network.neutron [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Updated VIF entry in instance network info cache for port 88de6684-404d-48bc-85c9-214ab6d77e42. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1083.119521] env[61985]: DEBUG nova.network.neutron [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Updating instance_info_cache with network_info: [{"id": "88de6684-404d-48bc-85c9-214ab6d77e42", "address": "fa:16:3e:93:90:1b", "network": {"id": "3e64e59c-d2d2-4d75-87a7-e63c5857c6a5", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1971303841-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2504537f92814207ab6453b6aea4f1fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88de6684-40", "ovs_interfaceid": "88de6684-404d-48bc-85c9-214ab6d77e42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.205688] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.242596] env[61985]: INFO nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating resource usage from migration ad0c29e8-fc04-41ba-a6ea-343ca3e29a1a [ 1083.247604] env[61985]: DEBUG nova.compute.manager [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1083.276451] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936478, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.280197] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 59e742d9-c51a-47dd-9ef2-2f070eae94c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.280352] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 073a7668-39e6-480d-9350-835a0282b456 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.280477] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 5f2aa808-8d1e-471e-9ef7-0f91590ec546 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.280602] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 44361f7b-4609-476f-b4a9-58a7851e6e92 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.280730] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 5aedbb9e-8bd1-499a-8195-b3045f7a141f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.280876] env[61985]: WARNING nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8c4a1fe4-4ded-4064-8261-24826181a68c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1083.280990] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 036fae78-2af0-49f1-85fa-245c29f35ebc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.281124] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 87006fbf-ea90-4d9a-88af-001de424ac14 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.281248] env[61985]: WARNING nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8eea60eb-93cf-4701-9ba3-0805691ca016 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1083.281370] env[61985]: WARNING nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance decfe088-90d6-49f9-9eae-7aa461219eaf is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1083.281483] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 9079e605-e865-4eb1-9857-03769e4b5ba6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.281595] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance acebe0a7-62d4-48d4-a1f7-5df0fa465753 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.281715] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8cfc2489-b3b9-463d-9ad5-0cd73920ed58 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.324464] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936487, 'name': CreateVM_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.400684] env[61985]: DEBUG oslo_vmware.api [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936484, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.498057] env[61985]: INFO nova.compute.manager [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Took 32.61 seconds to build instance. [ 1083.622608] env[61985]: DEBUG oslo_concurrency.lockutils [req-b694091e-d0b0-434f-85e7-4ad3ff690e09 req-c8657bd4-848e-4a34-aa2d-5308a92e5738 service nova] Releasing lock "refresh_cache-acebe0a7-62d4-48d4-a1f7-5df0fa465753" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.756375] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936478, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.88742} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.756375] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/173db284-6b8b-4355-bdde-6a9ba3be888e/173db284-6b8b-4355-bdde-6a9ba3be888e.vmdk to [datastore2] 87006fbf-ea90-4d9a-88af-001de424ac14/87006fbf-ea90-4d9a-88af-001de424ac14.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1083.757207] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ed4039-542e-4b22-b37f-ac8da2f384b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.784156] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 87006fbf-ea90-4d9a-88af-001de424ac14/87006fbf-ea90-4d9a-88af-001de424ac14.vmdk or device None with type streamOptimized {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.785557] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 926a2b59-8614-4e71-9d94-fd9feefdef46 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1083.786828] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cd71a74-ca2b-4f34-8115-19e3f15118a3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.803178] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1083.803349] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Migration ad0c29e8-fc04-41ba-a6ea-343ca3e29a1a is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1083.803461] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 885a36d9-24c6-407a-9d6b-01cd97e90674 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1083.803703] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1083.803819] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1083.812533] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1083.812533] env[61985]: value = "task-936488" [ 1083.812533] env[61985]: _type = "Task" [ 1083.812533] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.828429] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936487, 'name': CreateVM_Task, 'duration_secs': 0.667988} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.829294] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936488, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.829294] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1083.829826] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.829964] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.830302] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1083.830560] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c254569-6662-4166-963f-8bd596733b09 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.837150] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for the task: (returnval){ [ 1083.837150] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524b62b4-26db-fbf4-ddbf-1230478cdffd" [ 1083.837150] env[61985]: _type = "Task" [ 1083.837150] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.847020] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524b62b4-26db-fbf4-ddbf-1230478cdffd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.900312] env[61985]: DEBUG oslo_vmware.api [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Task: {'id': task-936484, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.279178} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.900580] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1083.900768] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1083.900954] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1083.901156] env[61985]: INFO nova.compute.manager [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Took 2.72 seconds to destroy the instance on the hypervisor. [ 1083.901411] env[61985]: DEBUG oslo.service.loopingcall [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.901601] env[61985]: DEBUG nova.compute.manager [-] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1083.901693] env[61985]: DEBUG nova.network.neutron [-] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1083.999755] env[61985]: DEBUG oslo_concurrency.lockutils [None req-77046d3f-99ce-4ab6-8a91-c4d60c14876b tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.133s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.014919] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c93c08-6568-439d-b8ed-3b062e98d374 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.021675] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f45fea-af45-4ad1-991c-fc172b2b3aa4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.055893] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c2deed-60ba-442e-8742-58c4b036457f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.062855] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984b34b6-e53a-4eb0-be13-816a7d6961e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.076386] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.208312] env[61985]: DEBUG nova.compute.manager [req-2bddc9eb-9552-43e7-af23-f770d01ef429 req-b92420d5-40f5-40a4-a794-d0f120688ea2 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Received event network-vif-deleted-7075efce-177a-4127-9c2f-7e3a9d1e92ad {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1084.208441] env[61985]: INFO nova.compute.manager [req-2bddc9eb-9552-43e7-af23-f770d01ef429 req-b92420d5-40f5-40a4-a794-d0f120688ea2 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Neutron deleted interface 7075efce-177a-4127-9c2f-7e3a9d1e92ad; detaching it from the instance and deleting it from the info cache [ 1084.208627] env[61985]: DEBUG nova.network.neutron [req-2bddc9eb-9552-43e7-af23-f770d01ef429 req-b92420d5-40f5-40a4-a794-d0f120688ea2 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.286264] env[61985]: DEBUG nova.compute.manager [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1084.323184] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936488, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.332429] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1084.332576] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1084.332722] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1084.332915] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1084.333086] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1084.333247] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1084.333465] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1084.333635] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1084.333817] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1084.333988] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1084.334188] env[61985]: DEBUG nova.virt.hardware [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1084.335057] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eabd7ca-d55f-408b-a616-3abb2b9a9f87 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.350776] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23953a32-3886-46d9-8928-3bdbe874e060 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.354758] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524b62b4-26db-fbf4-ddbf-1230478cdffd, 'name': SearchDatastore_Task, 'duration_secs': 0.010114} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.355421] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.356228] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1084.356228] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.356228] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.356373] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1084.357622] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4adf3ee3-6f04-4550-a6bc-c838bb94227c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.370113] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1084.370273] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1084.371042] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-677d96e5-783f-4f4c-960f-56668361053a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.376603] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for the task: (returnval){ [ 1084.376603] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a83807-7dba-ddda-26f7-2354c47c4596" [ 1084.376603] env[61985]: _type = "Task" [ 1084.376603] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.384590] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a83807-7dba-ddda-26f7-2354c47c4596, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.579611] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.628641] env[61985]: DEBUG nova.network.neutron [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Successfully updated port: e1f6f60d-06ee-469d-bf35-a3af36741d55 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1084.671914] env[61985]: DEBUG nova.compute.manager [req-ca20d11f-7dff-42f8-9dd5-0dd5e6686e43 req-dc9380d9-96ae-4bd2-8e7a-9ab3311dce0f service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Received event network-vif-plugged-e1f6f60d-06ee-469d-bf35-a3af36741d55 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1084.672196] env[61985]: DEBUG oslo_concurrency.lockutils [req-ca20d11f-7dff-42f8-9dd5-0dd5e6686e43 req-dc9380d9-96ae-4bd2-8e7a-9ab3311dce0f service nova] Acquiring lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.672434] env[61985]: DEBUG oslo_concurrency.lockutils [req-ca20d11f-7dff-42f8-9dd5-0dd5e6686e43 req-dc9380d9-96ae-4bd2-8e7a-9ab3311dce0f service nova] Lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.672617] env[61985]: DEBUG oslo_concurrency.lockutils [req-ca20d11f-7dff-42f8-9dd5-0dd5e6686e43 req-dc9380d9-96ae-4bd2-8e7a-9ab3311dce0f service nova] Lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.672793] env[61985]: DEBUG nova.compute.manager [req-ca20d11f-7dff-42f8-9dd5-0dd5e6686e43 req-dc9380d9-96ae-4bd2-8e7a-9ab3311dce0f service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] No waiting events found dispatching network-vif-plugged-e1f6f60d-06ee-469d-bf35-a3af36741d55 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1084.672963] env[61985]: WARNING nova.compute.manager [req-ca20d11f-7dff-42f8-9dd5-0dd5e6686e43 req-dc9380d9-96ae-4bd2-8e7a-9ab3311dce0f service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Received unexpected event network-vif-plugged-e1f6f60d-06ee-469d-bf35-a3af36741d55 for instance with vm_state building and task_state spawning. [ 1084.673276] env[61985]: DEBUG nova.network.neutron [-] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.711203] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a34c17c3-d5fb-4ca2-ad0c-88ad2251d2d9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.721065] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e3b22bf-01ff-4ceb-bef2-8b08f4ad23c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.750350] env[61985]: DEBUG nova.compute.manager [req-2bddc9eb-9552-43e7-af23-f770d01ef429 req-b92420d5-40f5-40a4-a794-d0f120688ea2 service nova] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Detach interface failed, port_id=7075efce-177a-4127-9c2f-7e3a9d1e92ad, reason: Instance 5f2aa808-8d1e-471e-9ef7-0f91590ec546 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1084.824688] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936488, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.887323] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a83807-7dba-ddda-26f7-2354c47c4596, 'name': SearchDatastore_Task, 'duration_secs': 0.010154} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.888358] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6030e4c-92e8-4bcc-b7ba-90c7348cbc45 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.894095] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for the task: (returnval){ [ 1084.894095] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52659cf3-a631-5ef0-d117-d724978f2c20" [ 1084.894095] env[61985]: _type = "Task" [ 1084.894095] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.903914] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52659cf3-a631-5ef0-d117-d724978f2c20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.084938] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1085.085245] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.859s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.085572] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.253s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.085785] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.088429] env[61985]: DEBUG oslo_concurrency.lockutils [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.255s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.088794] env[61985]: DEBUG oslo_concurrency.lockutils [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.091537] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.182s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.093945] env[61985]: INFO nova.compute.claims [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1085.121887] env[61985]: INFO nova.scheduler.client.report [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Deleted allocations for instance 8eea60eb-93cf-4701-9ba3-0805691ca016 [ 1085.125909] env[61985]: INFO nova.scheduler.client.report [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted allocations for instance decfe088-90d6-49f9-9eae-7aa461219eaf [ 1085.131230] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "refresh_cache-8cfc2489-b3b9-463d-9ad5-0cd73920ed58" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.133486] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "refresh_cache-8cfc2489-b3b9-463d-9ad5-0cd73920ed58" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.133486] env[61985]: DEBUG nova.network.neutron [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1085.177250] env[61985]: INFO nova.compute.manager [-] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Took 1.27 seconds to deallocate network for instance. [ 1085.323979] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936488, 'name': ReconfigVM_Task, 'duration_secs': 1.137765} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.324381] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 87006fbf-ea90-4d9a-88af-001de424ac14/87006fbf-ea90-4d9a-88af-001de424ac14.vmdk or device None with type streamOptimized {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1085.324931] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aaaa0155-f4b7-41b3-9052-a110daf55258 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.330593] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1085.330593] env[61985]: value = "task-936489" [ 1085.330593] env[61985]: _type = "Task" [ 1085.330593] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.338297] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936489, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.404472] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52659cf3-a631-5ef0-d117-d724978f2c20, 'name': SearchDatastore_Task, 'duration_secs': 0.010767} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.404689] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.404951] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] acebe0a7-62d4-48d4-a1f7-5df0fa465753/acebe0a7-62d4-48d4-a1f7-5df0fa465753.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1085.405221] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-193a833c-cf5b-4a04-9def-1ad542b74d7d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.411290] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for the task: (returnval){ [ 1085.411290] env[61985]: value = "task-936490" [ 1085.411290] env[61985]: _type = "Task" [ 1085.411290] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.418846] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936490, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.641833] env[61985]: DEBUG oslo_concurrency.lockutils [None req-7ae95725-e2c6-4db6-9849-5a174426fe06 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "decfe088-90d6-49f9-9eae-7aa461219eaf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.911s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.645678] env[61985]: DEBUG oslo_concurrency.lockutils [None req-69775ead-8fe4-48f4-ac28-cc89a7200695 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "8eea60eb-93cf-4701-9ba3-0805691ca016" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.836s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.682184] env[61985]: DEBUG nova.network.neutron [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1085.682184] env[61985]: DEBUG oslo_concurrency.lockutils [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.841735] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936489, 'name': Rename_Task, 'duration_secs': 0.131472} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.842207] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1085.842628] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e381444-f3cb-425c-9748-62a9cb47faf4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.849849] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1085.849849] env[61985]: value = "task-936491" [ 1085.849849] env[61985]: _type = "Task" [ 1085.849849] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.858070] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936491, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.865296] env[61985]: DEBUG nova.network.neutron [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Updating instance_info_cache with network_info: [{"id": "e1f6f60d-06ee-469d-bf35-a3af36741d55", "address": "fa:16:3e:08:4c:5d", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1f6f60d-06", "ovs_interfaceid": "e1f6f60d-06ee-469d-bf35-a3af36741d55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.922751] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936490, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441404} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.923105] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] acebe0a7-62d4-48d4-a1f7-5df0fa465753/acebe0a7-62d4-48d4-a1f7-5df0fa465753.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1085.923404] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.923715] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f62280f-ec3e-41c9-966b-f13b27fbb8ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.930576] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for the task: (returnval){ [ 1085.930576] env[61985]: value = "task-936492" [ 1085.930576] env[61985]: _type = "Task" [ 1085.930576] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.944307] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936492, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.238041] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.238326] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.239060] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.239060] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.239060] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.244381] env[61985]: DEBUG nova.compute.manager [req-33623e59-ebcf-4179-93d5-fcd9010e4600 req-a4202ad4-570e-491a-909a-73e6b0c35871 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Received event network-changed-1a081dc2-f497-4708-a3af-0b373d04cb85 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1086.244612] env[61985]: DEBUG nova.compute.manager [req-33623e59-ebcf-4179-93d5-fcd9010e4600 req-a4202ad4-570e-491a-909a-73e6b0c35871 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Refreshing instance network info cache due to event network-changed-1a081dc2-f497-4708-a3af-0b373d04cb85. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1086.244793] env[61985]: DEBUG oslo_concurrency.lockutils [req-33623e59-ebcf-4179-93d5-fcd9010e4600 req-a4202ad4-570e-491a-909a-73e6b0c35871 service nova] Acquiring lock "refresh_cache-9079e605-e865-4eb1-9857-03769e4b5ba6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.245111] env[61985]: DEBUG oslo_concurrency.lockutils [req-33623e59-ebcf-4179-93d5-fcd9010e4600 req-a4202ad4-570e-491a-909a-73e6b0c35871 service nova] Acquired lock "refresh_cache-9079e605-e865-4eb1-9857-03769e4b5ba6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.245435] env[61985]: DEBUG nova.network.neutron [req-33623e59-ebcf-4179-93d5-fcd9010e4600 req-a4202ad4-570e-491a-909a-73e6b0c35871 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Refreshing network info cache for port 1a081dc2-f497-4708-a3af-0b373d04cb85 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1086.247560] env[61985]: INFO nova.compute.manager [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Terminating instance [ 1086.250217] env[61985]: DEBUG nova.compute.manager [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1086.250419] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1086.251736] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967a166a-b608-4840-a129-adc3f670cfca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.263653] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1086.263809] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a65fdbe1-dcf4-4fcf-b087-405684875412 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.271829] env[61985]: DEBUG oslo_vmware.api [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1086.271829] env[61985]: value = "task-936493" [ 1086.271829] env[61985]: _type = "Task" [ 1086.271829] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.280051] env[61985]: DEBUG oslo_vmware.api [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936493, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.312287] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd87ab0-f250-4b92-866d-92f2ef6302c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.320532] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439025a6-7dd5-46b7-9e5d-050866c9caa4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.356055] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe96cb6-6d2b-41c3-a5bf-2bcf3603e34f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.363949] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936491, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.366981] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1abd6a77-2296-4c92-8080-0031007583d8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.370862] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "refresh_cache-8cfc2489-b3b9-463d-9ad5-0cd73920ed58" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.371175] env[61985]: DEBUG nova.compute.manager [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Instance network_info: |[{"id": "e1f6f60d-06ee-469d-bf35-a3af36741d55", "address": "fa:16:3e:08:4c:5d", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1f6f60d-06", "ovs_interfaceid": "e1f6f60d-06ee-469d-bf35-a3af36741d55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1086.371558] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:4c:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7908211b-df93-467b-87a8-3c3d29b03de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e1f6f60d-06ee-469d-bf35-a3af36741d55', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1086.378949] env[61985]: DEBUG oslo.service.loopingcall [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1086.379547] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1086.379782] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c799ce92-f568-4aa3-ab43-9e2ddb89fed8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.404175] env[61985]: DEBUG nova.compute.provider_tree [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.410387] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1086.410387] env[61985]: value = "task-936494" [ 1086.410387] env[61985]: _type = "Task" [ 1086.410387] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.418924] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936494, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.439540] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936492, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060888} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.439804] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.440583] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84fe224-cc30-4ef3-bfcd-c22adc7b1811 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.462833] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] acebe0a7-62d4-48d4-a1f7-5df0fa465753/acebe0a7-62d4-48d4-a1f7-5df0fa465753.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.463196] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cc96306-4d93-4247-af34-8e21968bbbb5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.482164] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for the task: (returnval){ [ 1086.482164] env[61985]: value = "task-936495" [ 1086.482164] env[61985]: _type = "Task" [ 1086.482164] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.491855] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936495, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.702417] env[61985]: DEBUG nova.compute.manager [req-980a02ba-bd1b-4723-91cb-cef251af8cf3 req-7b4c178f-10a1-49db-8a8d-b419c40aec6a service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Received event network-changed-e1f6f60d-06ee-469d-bf35-a3af36741d55 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1086.702709] env[61985]: DEBUG nova.compute.manager [req-980a02ba-bd1b-4723-91cb-cef251af8cf3 req-7b4c178f-10a1-49db-8a8d-b419c40aec6a service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Refreshing instance network info cache due to event network-changed-e1f6f60d-06ee-469d-bf35-a3af36741d55. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1086.702955] env[61985]: DEBUG oslo_concurrency.lockutils [req-980a02ba-bd1b-4723-91cb-cef251af8cf3 req-7b4c178f-10a1-49db-8a8d-b419c40aec6a service nova] Acquiring lock "refresh_cache-8cfc2489-b3b9-463d-9ad5-0cd73920ed58" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.703429] env[61985]: DEBUG oslo_concurrency.lockutils [req-980a02ba-bd1b-4723-91cb-cef251af8cf3 req-7b4c178f-10a1-49db-8a8d-b419c40aec6a service nova] Acquired lock "refresh_cache-8cfc2489-b3b9-463d-9ad5-0cd73920ed58" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.703429] env[61985]: DEBUG nova.network.neutron [req-980a02ba-bd1b-4723-91cb-cef251af8cf3 req-7b4c178f-10a1-49db-8a8d-b419c40aec6a service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Refreshing network info cache for port e1f6f60d-06ee-469d-bf35-a3af36741d55 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1086.790121] env[61985]: DEBUG oslo_vmware.api [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936493, 'name': PowerOffVM_Task, 'duration_secs': 0.343303} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.791209] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1086.791326] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1086.791595] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dbea4624-d6c9-4ab2-886b-19ae3166bd88 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.864402] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936491, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.907444] env[61985]: DEBUG nova.scheduler.client.report [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1086.922541] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936494, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.997742] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936495, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.999480] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1086.999692] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1086.999889] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleting the datastore file [datastore2] 5aedbb9e-8bd1-499a-8195-b3045f7a141f {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1087.000897] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ec539b9-3e7a-408b-8b05-4cd32c40be9b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.008103] env[61985]: DEBUG oslo_vmware.api [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1087.008103] env[61985]: value = "task-936497" [ 1087.008103] env[61985]: _type = "Task" [ 1087.008103] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.017008] env[61985]: DEBUG oslo_vmware.api [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.148312] env[61985]: DEBUG nova.network.neutron [req-33623e59-ebcf-4179-93d5-fcd9010e4600 req-a4202ad4-570e-491a-909a-73e6b0c35871 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Updated VIF entry in instance network info cache for port 1a081dc2-f497-4708-a3af-0b373d04cb85. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1087.148312] env[61985]: DEBUG nova.network.neutron [req-33623e59-ebcf-4179-93d5-fcd9010e4600 req-a4202ad4-570e-491a-909a-73e6b0c35871 service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Updating instance_info_cache with network_info: [{"id": "1a081dc2-f497-4708-a3af-0b373d04cb85", "address": "fa:16:3e:5c:a1:b1", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.247", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a081dc2-f4", "ovs_interfaceid": "1a081dc2-f497-4708-a3af-0b373d04cb85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.363408] env[61985]: DEBUG oslo_vmware.api [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936491, 'name': PowerOnVM_Task, 'duration_secs': 1.140274} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.363709] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1087.417564] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.326s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.418191] env[61985]: DEBUG nova.compute.manager [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1087.425074] env[61985]: DEBUG oslo_concurrency.lockutils [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.251s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.425074] env[61985]: DEBUG oslo_concurrency.lockutils [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.427219] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.564s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.429129] env[61985]: INFO nova.compute.claims [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.431054] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936494, 'name': CreateVM_Task, 'duration_secs': 0.615453} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.432271] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1087.433328] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.433328] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.433328] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1087.433328] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fd57411-2a7d-4147-afef-46e7d64a4a32 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.441153] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1087.441153] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521f574f-588a-461d-6674-2df92e8edd17" [ 1087.441153] env[61985]: _type = "Task" [ 1087.441153] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.449721] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521f574f-588a-461d-6674-2df92e8edd17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.475133] env[61985]: INFO nova.scheduler.client.report [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted allocations for instance 8c4a1fe4-4ded-4064-8261-24826181a68c [ 1087.480806] env[61985]: DEBUG nova.network.neutron [req-980a02ba-bd1b-4723-91cb-cef251af8cf3 req-7b4c178f-10a1-49db-8a8d-b419c40aec6a service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Updated VIF entry in instance network info cache for port e1f6f60d-06ee-469d-bf35-a3af36741d55. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1087.481260] env[61985]: DEBUG nova.network.neutron [req-980a02ba-bd1b-4723-91cb-cef251af8cf3 req-7b4c178f-10a1-49db-8a8d-b419c40aec6a service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Updating instance_info_cache with network_info: [{"id": "e1f6f60d-06ee-469d-bf35-a3af36741d55", "address": "fa:16:3e:08:4c:5d", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1f6f60d-06", "ovs_interfaceid": "e1f6f60d-06ee-469d-bf35-a3af36741d55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.493839] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936495, 'name': ReconfigVM_Task, 'duration_secs': 0.555542} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.494196] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Reconfigured VM instance instance-00000063 to attach disk [datastore1] acebe0a7-62d4-48d4-a1f7-5df0fa465753/acebe0a7-62d4-48d4-a1f7-5df0fa465753.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.495341] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0b807e3-1bc7-4fc7-be5b-b700b453a493 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.501644] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for the task: (returnval){ [ 1087.501644] env[61985]: value = "task-936498" [ 1087.501644] env[61985]: _type = "Task" [ 1087.501644] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.505836] env[61985]: DEBUG nova.compute.manager [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1087.506615] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5147cd38-de19-4b23-b2e4-8baf574b0481 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.516773] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936498, 'name': Rename_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.523809] env[61985]: DEBUG oslo_vmware.api [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171392} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.524328] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1087.524577] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1087.524734] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1087.524913] env[61985]: INFO nova.compute.manager [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Took 1.27 seconds to destroy the instance on the hypervisor. [ 1087.525207] env[61985]: DEBUG oslo.service.loopingcall [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1087.525414] env[61985]: DEBUG nova.compute.manager [-] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1087.525594] env[61985]: DEBUG nova.network.neutron [-] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1087.651600] env[61985]: DEBUG oslo_concurrency.lockutils [req-33623e59-ebcf-4179-93d5-fcd9010e4600 req-a4202ad4-570e-491a-909a-73e6b0c35871 service nova] Releasing lock "refresh_cache-9079e605-e865-4eb1-9857-03769e4b5ba6" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.933879] env[61985]: DEBUG nova.compute.utils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1087.937474] env[61985]: DEBUG nova.compute.manager [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1087.937701] env[61985]: DEBUG nova.network.neutron [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1087.953130] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521f574f-588a-461d-6674-2df92e8edd17, 'name': SearchDatastore_Task, 'duration_secs': 0.020445} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.953537] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.953718] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.953969] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.954131] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.954332] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.956884] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca939a0e-e4ef-4953-83ad-e5bdc56f9b40 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.966577] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.966577] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1087.967105] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24edf041-cf10-4bd1-b6e3-8316da62e469 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.974038] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1087.974038] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5246d25a-6316-e4d2-1d7f-b3f2f68a6bf8" [ 1087.974038] env[61985]: _type = "Task" [ 1087.974038] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.982593] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5246d25a-6316-e4d2-1d7f-b3f2f68a6bf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.987388] env[61985]: DEBUG oslo_concurrency.lockutils [None req-60ad49c2-5544-4c84-947f-2bfc053e87a3 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "8c4a1fe4-4ded-4064-8261-24826181a68c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.918s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.989446] env[61985]: DEBUG nova.policy [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '344f7f484f4240cd9dc32b200b697e12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfc57e6393c94545be32165d41230db3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1087.990974] env[61985]: DEBUG oslo_concurrency.lockutils [req-980a02ba-bd1b-4723-91cb-cef251af8cf3 req-7b4c178f-10a1-49db-8a8d-b419c40aec6a service nova] Releasing lock "refresh_cache-8cfc2489-b3b9-463d-9ad5-0cd73920ed58" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.011946] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936498, 'name': Rename_Task, 'duration_secs': 0.358628} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.012095] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1088.012308] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f42858a7-3bd1-437f-a2a4-3ee80d11ba46 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.020171] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for the task: (returnval){ [ 1088.020171] env[61985]: value = "task-936499" [ 1088.020171] env[61985]: _type = "Task" [ 1088.020171] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.034088] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936499, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.037067] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0d90d3c6-aad3-4a2b-ad47-22fb6e4ccd8e tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "87006fbf-ea90-4d9a-88af-001de424ac14" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 38.523s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.438457] env[61985]: DEBUG nova.compute.manager [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1088.457247] env[61985]: DEBUG nova.network.neutron [-] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.487056] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5246d25a-6316-e4d2-1d7f-b3f2f68a6bf8, 'name': SearchDatastore_Task, 'duration_secs': 0.011931} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.491246] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-057012e0-9b34-43a2-9bee-28c6fcf6bb6e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.497742] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1088.497742] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fa7c87-98be-9878-748e-582ecbf85a78" [ 1088.497742] env[61985]: _type = "Task" [ 1088.497742] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.513149] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fa7c87-98be-9878-748e-582ecbf85a78, 'name': SearchDatastore_Task, 'duration_secs': 0.010547} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.513149] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.513149] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 8cfc2489-b3b9-463d-9ad5-0cd73920ed58/8cfc2489-b3b9-463d-9ad5-0cd73920ed58.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1088.513320] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5572d9b8-2491-4e69-9177-b8844b67298b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.516887] env[61985]: DEBUG nova.network.neutron [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Successfully created port: 4249c70c-abfc-4d75-b161-1bfda2047dd6 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1088.528341] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1088.528341] env[61985]: value = "task-936500" [ 1088.528341] env[61985]: _type = "Task" [ 1088.528341] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.537378] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936499, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.544859] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936500, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.556594] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.556854] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.684650] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab265be3-633d-4d9e-b6aa-e9093d256568 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.692851] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da427959-622d-42d1-a696-701fa92584fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.725300] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a181f3c-e226-453b-b062-3b2186c6814b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.733250] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e13c50f-a3de-44b1-88bd-80e76e7ca41c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.750119] env[61985]: DEBUG nova.compute.provider_tree [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.885101] env[61985]: DEBUG nova.compute.manager [req-10d4729b-eacd-4316-9001-e329b2158499 req-3ff53d0d-fac0-46e8-b988-04853ddf3cdb service nova] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Received event network-vif-deleted-36a3c11a-0e39-4b68-ada2-17d3d6a82a56 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1088.963364] env[61985]: INFO nova.compute.manager [-] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Took 1.44 seconds to deallocate network for instance. [ 1089.033757] env[61985]: DEBUG oslo_vmware.api [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936499, 'name': PowerOnVM_Task, 'duration_secs': 0.725555} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.039327] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1089.039327] env[61985]: INFO nova.compute.manager [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Took 8.68 seconds to spawn the instance on the hypervisor. [ 1089.039327] env[61985]: DEBUG nova.compute.manager [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1089.039327] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0d185a-d0ce-487e-a28f-7a7d9b19e17b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.052953] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936500, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.058978] env[61985]: DEBUG nova.compute.manager [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1089.254398] env[61985]: DEBUG nova.scheduler.client.report [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1089.451157] env[61985]: DEBUG nova.compute.manager [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1089.470757] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.480812] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1089.481077] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1089.481277] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1089.481508] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1089.481753] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1089.481933] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1089.482167] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1089.482340] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1089.482515] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1089.482686] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1089.482868] env[61985]: DEBUG nova.virt.hardware [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1089.483734] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f233009-454b-4511-9f02-0e63efad1a64 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.492041] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f2f982-298c-4b96-a90a-caccd0310444 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.541973] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936500, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521938} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.542260] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 8cfc2489-b3b9-463d-9ad5-0cd73920ed58/8cfc2489-b3b9-463d-9ad5-0cd73920ed58.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1089.542480] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1089.542728] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1eadc34d-e197-430c-8b0a-87f0faeeb65a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.548954] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1089.548954] env[61985]: value = "task-936501" [ 1089.548954] env[61985]: _type = "Task" [ 1089.548954] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.560323] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936501, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.564105] env[61985]: INFO nova.compute.manager [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Took 15.28 seconds to build instance. [ 1089.584797] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.759710] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.760236] env[61985]: DEBUG nova.compute.manager [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1089.763183] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.558s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.063586] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936501, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089498} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.063586] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1090.063586] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a4aadd-5890-4058-93bb-e57fc61a3a4c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.081014] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e15e8f1e-267d-4278-a1df-c74896603be1 tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.816s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.092674] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 8cfc2489-b3b9-463d-9ad5-0cd73920ed58/8cfc2489-b3b9-463d-9ad5-0cd73920ed58.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1090.093251] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7862053-40f2-4e4a-a54b-7f239b9943a8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.118199] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1090.118199] env[61985]: value = "task-936502" [ 1090.118199] env[61985]: _type = "Task" [ 1090.118199] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.127934] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936502, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.266692] env[61985]: DEBUG nova.compute.utils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1090.273627] env[61985]: INFO nova.compute.claims [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1090.279693] env[61985]: DEBUG nova.compute.manager [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1090.279900] env[61985]: DEBUG nova.network.neutron [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1090.295733] env[61985]: DEBUG nova.compute.manager [req-0daaeec3-ef7d-4642-9088-8b48bbcb1186 req-b03749f0-d1b5-4797-a8a5-92785c3fd95e service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Received event network-vif-plugged-4249c70c-abfc-4d75-b161-1bfda2047dd6 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1090.296491] env[61985]: DEBUG oslo_concurrency.lockutils [req-0daaeec3-ef7d-4642-9088-8b48bbcb1186 req-b03749f0-d1b5-4797-a8a5-92785c3fd95e service nova] Acquiring lock "926a2b59-8614-4e71-9d94-fd9feefdef46-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.298249] env[61985]: DEBUG oslo_concurrency.lockutils [req-0daaeec3-ef7d-4642-9088-8b48bbcb1186 req-b03749f0-d1b5-4797-a8a5-92785c3fd95e service nova] Lock "926a2b59-8614-4e71-9d94-fd9feefdef46-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.298249] env[61985]: DEBUG oslo_concurrency.lockutils [req-0daaeec3-ef7d-4642-9088-8b48bbcb1186 req-b03749f0-d1b5-4797-a8a5-92785c3fd95e service nova] Lock "926a2b59-8614-4e71-9d94-fd9feefdef46-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.298249] env[61985]: DEBUG nova.compute.manager [req-0daaeec3-ef7d-4642-9088-8b48bbcb1186 req-b03749f0-d1b5-4797-a8a5-92785c3fd95e service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] No waiting events found dispatching network-vif-plugged-4249c70c-abfc-4d75-b161-1bfda2047dd6 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1090.298249] env[61985]: WARNING nova.compute.manager [req-0daaeec3-ef7d-4642-9088-8b48bbcb1186 req-b03749f0-d1b5-4797-a8a5-92785c3fd95e service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Received unexpected event network-vif-plugged-4249c70c-abfc-4d75-b161-1bfda2047dd6 for instance with vm_state building and task_state spawning. [ 1090.339714] env[61985]: DEBUG nova.policy [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b489c06ceada46bea5190b44dafd777a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e7e22f7d7b7b407ebb2d00e174d8a42c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1090.405469] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b955e05-17ea-4da7-9047-950371bd8b93 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.411140] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-76f55fa0-1187-407b-b216-2242b97c8b5c tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Suspending the VM {{(pid=61985) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1090.416148] env[61985]: DEBUG nova.network.neutron [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Successfully updated port: 4249c70c-abfc-4d75-b161-1bfda2047dd6 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1090.416148] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-c6ed7bfc-e2d2-48b7-a505-9e3c77349c9e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.421776] env[61985]: DEBUG oslo_vmware.api [None req-76f55fa0-1187-407b-b216-2242b97c8b5c tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1090.421776] env[61985]: value = "task-936503" [ 1090.421776] env[61985]: _type = "Task" [ 1090.421776] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.432595] env[61985]: DEBUG oslo_vmware.api [None req-76f55fa0-1187-407b-b216-2242b97c8b5c tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936503, 'name': SuspendVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.633183] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936502, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.648671] env[61985]: DEBUG nova.network.neutron [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Successfully created port: 03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1090.695984] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquiring lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.695984] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.695984] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquiring lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.695984] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.695984] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.699442] env[61985]: INFO nova.compute.manager [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Terminating instance [ 1090.702140] env[61985]: DEBUG nova.compute.manager [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1090.702587] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1090.704031] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2cbd5eb-2cac-499a-8354-e72d2fe488d4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.712902] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1090.714277] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-190d5089-1fba-4b20-8751-36453d84cd5e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.722018] env[61985]: DEBUG oslo_vmware.api [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for the task: (returnval){ [ 1090.722018] env[61985]: value = "task-936504" [ 1090.722018] env[61985]: _type = "Task" [ 1090.722018] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.727917] env[61985]: DEBUG oslo_vmware.api [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936504, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.783416] env[61985]: DEBUG nova.compute.manager [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1090.794437] env[61985]: INFO nova.compute.resource_tracker [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating resource usage from migration ad0c29e8-fc04-41ba-a6ea-343ca3e29a1a [ 1090.915991] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "refresh_cache-926a2b59-8614-4e71-9d94-fd9feefdef46" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.916180] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "refresh_cache-926a2b59-8614-4e71-9d94-fd9feefdef46" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.916341] env[61985]: DEBUG nova.network.neutron [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1090.935467] env[61985]: DEBUG oslo_vmware.api [None req-76f55fa0-1187-407b-b216-2242b97c8b5c tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936503, 'name': SuspendVM_Task} progress is 62%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.048843] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deff5bbc-aa48-4d5d-aec3-051535aab48b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.061034] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432bb42a-ac6e-45d7-9e0d-0c97282bf910 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.092598] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b5ec54-cebf-44e2-8e75-43e5a2698902 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.099961] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d309d8-05d2-412c-864d-e90b67cfa0f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.115332] env[61985]: DEBUG nova.compute.provider_tree [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.131121] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936502, 'name': ReconfigVM_Task, 'duration_secs': 0.988992} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.131477] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 8cfc2489-b3b9-463d-9ad5-0cd73920ed58/8cfc2489-b3b9-463d-9ad5-0cd73920ed58.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1091.132155] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a32ddfc-65d1-4977-bafb-8ffac3f183d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.138372] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1091.138372] env[61985]: value = "task-936505" [ 1091.138372] env[61985]: _type = "Task" [ 1091.138372] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.150296] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936505, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.231467] env[61985]: DEBUG oslo_vmware.api [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936504, 'name': PowerOffVM_Task, 'duration_secs': 0.411755} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.231467] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1091.231467] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1091.231467] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dce34740-ac5f-4459-8bb0-fd3068f6286f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.380833] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1091.381119] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1091.381322] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Deleting the datastore file [datastore1] acebe0a7-62d4-48d4-a1f7-5df0fa465753 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.381618] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-542cedea-4c12-4665-b8e9-e97c9b6d9cb4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.388611] env[61985]: DEBUG oslo_vmware.api [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for the task: (returnval){ [ 1091.388611] env[61985]: value = "task-936507" [ 1091.388611] env[61985]: _type = "Task" [ 1091.388611] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.396402] env[61985]: DEBUG oslo_vmware.api [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936507, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.433311] env[61985]: DEBUG oslo_vmware.api [None req-76f55fa0-1187-407b-b216-2242b97c8b5c tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936503, 'name': SuspendVM_Task, 'duration_secs': 0.929505} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.433602] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-76f55fa0-1187-407b-b216-2242b97c8b5c tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Suspended the VM {{(pid=61985) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1091.433796] env[61985]: DEBUG nova.compute.manager [None req-76f55fa0-1187-407b-b216-2242b97c8b5c tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1091.434571] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630775b8-bc6e-40dd-be5d-6eab436d1142 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.470056] env[61985]: DEBUG nova.network.neutron [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1091.618440] env[61985]: DEBUG nova.scheduler.client.report [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1091.648634] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936505, 'name': Rename_Task, 'duration_secs': 0.143876} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.648959] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1091.649349] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07eab2d1-cc77-415b-b6a0-1dea1babf32b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.657018] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1091.657018] env[61985]: value = "task-936508" [ 1091.657018] env[61985]: _type = "Task" [ 1091.657018] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.671750] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936508, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.689452] env[61985]: DEBUG nova.network.neutron [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Updating instance_info_cache with network_info: [{"id": "4249c70c-abfc-4d75-b161-1bfda2047dd6", "address": "fa:16:3e:35:ee:a6", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4249c70c-ab", "ovs_interfaceid": "4249c70c-abfc-4d75-b161-1bfda2047dd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.793396] env[61985]: DEBUG nova.compute.manager [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1091.822222] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.822490] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.822653] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.822888] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.823074] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.823236] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.823461] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.823629] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.823805] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.823977] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.824171] env[61985]: DEBUG nova.virt.hardware [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.825058] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73010109-3a39-4c77-bdb7-341db0fc79b6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.833947] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7ef1a7-5860-4aa6-99a9-70335f621e30 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.898267] env[61985]: DEBUG oslo_vmware.api [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Task: {'id': task-936507, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138187} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.898536] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.898765] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1091.898955] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1091.899175] env[61985]: INFO nova.compute.manager [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1091.899456] env[61985]: DEBUG oslo.service.loopingcall [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.899656] env[61985]: DEBUG nova.compute.manager [-] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.899754] env[61985]: DEBUG nova.network.neutron [-] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1092.127029] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.361s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.127029] env[61985]: INFO nova.compute.manager [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Migrating [ 1092.132541] env[61985]: DEBUG oslo_concurrency.lockutils [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.451s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.133253] env[61985]: DEBUG nova.objects.instance [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lazy-loading 'resources' on Instance uuid 5f2aa808-8d1e-471e-9ef7-0f91590ec546 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.165793] env[61985]: DEBUG nova.network.neutron [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Successfully updated port: 03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1092.173397] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936508, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.196032] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "refresh_cache-926a2b59-8614-4e71-9d94-fd9feefdef46" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.196032] env[61985]: DEBUG nova.compute.manager [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Instance network_info: |[{"id": "4249c70c-abfc-4d75-b161-1bfda2047dd6", "address": "fa:16:3e:35:ee:a6", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4249c70c-ab", "ovs_interfaceid": "4249c70c-abfc-4d75-b161-1bfda2047dd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1092.196032] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:ee:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bff6c3a1-cc80-46ca-86c0-6dbb029edddb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4249c70c-abfc-4d75-b161-1bfda2047dd6', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1092.203259] env[61985]: DEBUG oslo.service.loopingcall [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.203914] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1092.204285] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dabb721e-5ad1-47db-9084-96447db5116e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.225876] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1092.225876] env[61985]: value = "task-936509" [ 1092.225876] env[61985]: _type = "Task" [ 1092.225876] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.235389] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936509, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.262796] env[61985]: DEBUG nova.compute.manager [req-f4bfa97e-8d3f-4579-9661-596d48bcecdf req-581dde4a-7c88-427b-82d4-5806ab63d2cd service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received event network-vif-plugged-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1092.263058] env[61985]: DEBUG oslo_concurrency.lockutils [req-f4bfa97e-8d3f-4579-9661-596d48bcecdf req-581dde4a-7c88-427b-82d4-5806ab63d2cd service nova] Acquiring lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.263268] env[61985]: DEBUG oslo_concurrency.lockutils [req-f4bfa97e-8d3f-4579-9661-596d48bcecdf req-581dde4a-7c88-427b-82d4-5806ab63d2cd service nova] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.263472] env[61985]: DEBUG oslo_concurrency.lockutils [req-f4bfa97e-8d3f-4579-9661-596d48bcecdf req-581dde4a-7c88-427b-82d4-5806ab63d2cd service nova] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.263613] env[61985]: DEBUG nova.compute.manager [req-f4bfa97e-8d3f-4579-9661-596d48bcecdf req-581dde4a-7c88-427b-82d4-5806ab63d2cd service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] No waiting events found dispatching network-vif-plugged-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1092.263790] env[61985]: WARNING nova.compute.manager [req-f4bfa97e-8d3f-4579-9661-596d48bcecdf req-581dde4a-7c88-427b-82d4-5806ab63d2cd service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received unexpected event network-vif-plugged-03b6ccf4-e23e-472e-a02a-e72bd1686c90 for instance with vm_state building and task_state spawning. [ 1092.401671] env[61985]: DEBUG nova.compute.manager [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Received event network-changed-4249c70c-abfc-4d75-b161-1bfda2047dd6 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1092.401912] env[61985]: DEBUG nova.compute.manager [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Refreshing instance network info cache due to event network-changed-4249c70c-abfc-4d75-b161-1bfda2047dd6. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1092.402123] env[61985]: DEBUG oslo_concurrency.lockutils [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] Acquiring lock "refresh_cache-926a2b59-8614-4e71-9d94-fd9feefdef46" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.402287] env[61985]: DEBUG oslo_concurrency.lockutils [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] Acquired lock "refresh_cache-926a2b59-8614-4e71-9d94-fd9feefdef46" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.402442] env[61985]: DEBUG nova.network.neutron [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Refreshing network info cache for port 4249c70c-abfc-4d75-b161-1bfda2047dd6 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1092.651648] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.651887] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.651988] env[61985]: DEBUG nova.network.neutron [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1092.667567] env[61985]: DEBUG oslo_vmware.api [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936508, 'name': PowerOnVM_Task, 'duration_secs': 0.85375} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.670322] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1092.670517] env[61985]: INFO nova.compute.manager [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Took 8.38 seconds to spawn the instance on the hypervisor. [ 1092.670719] env[61985]: DEBUG nova.compute.manager [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1092.671216] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.671371] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.671535] env[61985]: DEBUG nova.network.neutron [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1092.673795] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e241c8a5-333d-4a50-9385-f9032079843f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.722040] env[61985]: DEBUG nova.network.neutron [-] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.737702] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936509, 'name': CreateVM_Task, 'duration_secs': 0.345928} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.740115] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1092.741552] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.741552] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.741849] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1092.742130] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84e820ab-9afa-4ad1-9259-67949cb4dc63 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.746946] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1092.746946] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a61c9a-2087-b771-dc77-b79369b46deb" [ 1092.746946] env[61985]: _type = "Task" [ 1092.746946] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.757946] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a61c9a-2087-b771-dc77-b79369b46deb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.847518] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e05a1d-4500-42b3-a0d6-0dcfe82d8b2a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.855324] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a799eeb-b771-40c9-8e8b-711e8331eb8b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.885431] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e97cfe-4fc2-450d-9cc2-235931b8abbb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.892518] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8af4d72-32bf-4e11-b142-2b682f0c8a01 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.907033] env[61985]: DEBUG nova.compute.provider_tree [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.073841] env[61985]: INFO nova.compute.manager [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Resuming [ 1093.074602] env[61985]: DEBUG nova.objects.instance [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lazy-loading 'flavor' on Instance uuid 87006fbf-ea90-4d9a-88af-001de424ac14 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.111824] env[61985]: DEBUG nova.network.neutron [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Updated VIF entry in instance network info cache for port 4249c70c-abfc-4d75-b161-1bfda2047dd6. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1093.112594] env[61985]: DEBUG nova.network.neutron [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Updating instance_info_cache with network_info: [{"id": "4249c70c-abfc-4d75-b161-1bfda2047dd6", "address": "fa:16:3e:35:ee:a6", "network": {"id": "a49e6c82-35d0-4620-9980-2dce0f1dc93e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1485776201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfc57e6393c94545be32165d41230db3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bff6c3a1-cc80-46ca-86c0-6dbb029edddb", "external-id": "nsx-vlan-transportzone-223", "segmentation_id": 223, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4249c70c-ab", "ovs_interfaceid": "4249c70c-abfc-4d75-b161-1bfda2047dd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.192075] env[61985]: INFO nova.compute.manager [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Took 16.63 seconds to build instance. [ 1093.203726] env[61985]: DEBUG nova.network.neutron [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1093.226985] env[61985]: INFO nova.compute.manager [-] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Took 1.33 seconds to deallocate network for instance. [ 1093.258326] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a61c9a-2087-b771-dc77-b79369b46deb, 'name': SearchDatastore_Task, 'duration_secs': 0.011029} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.258326] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.258326] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1093.258731] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.258731] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.258731] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1093.258980] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ee23fc9-0544-4cee-b865-031f83c48130 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.267065] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1093.267263] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1093.268049] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cedd1f44-f382-4f2f-b0e7-a8c483ffe479 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.276281] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1093.276281] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52636a5b-8c51-ca07-fb60-f6089696dbbb" [ 1093.276281] env[61985]: _type = "Task" [ 1093.276281] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.284355] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52636a5b-8c51-ca07-fb60-f6089696dbbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.346558] env[61985]: DEBUG nova.network.neutron [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating instance_info_cache with network_info: [{"id": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "address": "fa:16:3e:7c:35:40", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03b6ccf4-e2", "ovs_interfaceid": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.413470] env[61985]: DEBUG nova.scheduler.client.report [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.485132] env[61985]: DEBUG nova.network.neutron [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance_info_cache with network_info: [{"id": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "address": "fa:16:3e:e1:d0:1f", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64c525f-b3", "ovs_interfaceid": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.614870] env[61985]: DEBUG oslo_concurrency.lockutils [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] Releasing lock "refresh_cache-926a2b59-8614-4e71-9d94-fd9feefdef46" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.615175] env[61985]: DEBUG nova.compute.manager [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Received event network-vif-deleted-88de6684-404d-48bc-85c9-214ab6d77e42 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1093.615365] env[61985]: INFO nova.compute.manager [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Neutron deleted interface 88de6684-404d-48bc-85c9-214ab6d77e42; detaching it from the instance and deleting it from the info cache [ 1093.615563] env[61985]: DEBUG nova.network.neutron [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.693368] env[61985]: DEBUG oslo_concurrency.lockutils [None req-f29ba4f1-3315-412a-8e6d-0e978c3779c4 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.146s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.735233] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.787221] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52636a5b-8c51-ca07-fb60-f6089696dbbb, 'name': SearchDatastore_Task, 'duration_secs': 0.008657} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.788265] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e6f01b9-4b59-4abb-8fe4-b03e9913eacc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.795051] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1093.795051] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b2d1ec-8b60-d394-677c-b4408fed42ae" [ 1093.795051] env[61985]: _type = "Task" [ 1093.795051] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.804906] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b2d1ec-8b60-d394-677c-b4408fed42ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.809380] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "44361f7b-4609-476f-b4a9-58a7851e6e92" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.809380] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.809380] env[61985]: INFO nova.compute.manager [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Rebooting instance [ 1093.848857] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.849444] env[61985]: DEBUG nova.compute.manager [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Instance network_info: |[{"id": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "address": "fa:16:3e:7c:35:40", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03b6ccf4-e2", "ovs_interfaceid": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1093.850081] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:35:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e55c248-c504-4c7a-bbe9-f42cf417aee7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03b6ccf4-e23e-472e-a02a-e72bd1686c90', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1093.858801] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating folder: Project (e7e22f7d7b7b407ebb2d00e174d8a42c). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1093.861653] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7fe1aaad-7964-4e69-badf-04535bf8b001 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.870985] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Created folder: Project (e7e22f7d7b7b407ebb2d00e174d8a42c) in parent group-v211285. [ 1093.871143] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating folder: Instances. Parent ref: group-v211472. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1093.871431] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c315083c-3af0-4b04-a934-8c28379c39a8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.881743] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Created folder: Instances in parent group-v211472. [ 1093.881743] env[61985]: DEBUG oslo.service.loopingcall [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1093.881743] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1093.881743] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0623dd7d-18fe-4fe2-9612-ffb17b6236cd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.904901] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1093.904901] env[61985]: value = "task-936512" [ 1093.904901] env[61985]: _type = "Task" [ 1093.904901] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.912287] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936512, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.918157] env[61985]: DEBUG oslo_concurrency.lockutils [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.920226] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.450s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.920462] env[61985]: DEBUG nova.objects.instance [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lazy-loading 'resources' on Instance uuid 5aedbb9e-8bd1-499a-8195-b3045f7a141f {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.944256] env[61985]: INFO nova.scheduler.client.report [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Deleted allocations for instance 5f2aa808-8d1e-471e-9ef7-0f91590ec546 [ 1093.990136] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.082281] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.082479] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquired lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.082669] env[61985]: DEBUG nova.network.neutron [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1094.119075] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d63d745-4c11-4bab-8b18-e97c8a76fc0e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.127885] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da4b86b-3121-465b-9fe4-617cf3a6e7c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.155812] env[61985]: DEBUG nova.compute.manager [req-c5d3f053-b890-4447-8016-ce7d5b333d7f req-55327d28-0023-401d-a836-905e3b258dea service nova] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Detach interface failed, port_id=88de6684-404d-48bc-85c9-214ab6d77e42, reason: Instance acebe0a7-62d4-48d4-a1f7-5df0fa465753 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1094.229111] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11eb5cc5-dbb6-4358-838b-cd908bd93b05 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.235404] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d662ccbb-89ed-40ef-b61d-ea63c5baeefb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Suspending the VM {{(pid=61985) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1094.235680] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-f9e2fe42-2954-4447-ac37-c86861c3ee83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.240973] env[61985]: DEBUG oslo_vmware.api [None req-d662ccbb-89ed-40ef-b61d-ea63c5baeefb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1094.240973] env[61985]: value = "task-936513" [ 1094.240973] env[61985]: _type = "Task" [ 1094.240973] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.248160] env[61985]: DEBUG oslo_vmware.api [None req-d662ccbb-89ed-40ef-b61d-ea63c5baeefb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936513, 'name': SuspendVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.289399] env[61985]: DEBUG nova.compute.manager [req-d64ce9fc-d75c-42f7-b323-5d3f975b6f10 req-8aeb0800-44a4-460a-b15c-6a066147dd3f service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received event network-changed-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1094.289658] env[61985]: DEBUG nova.compute.manager [req-d64ce9fc-d75c-42f7-b323-5d3f975b6f10 req-8aeb0800-44a4-460a-b15c-6a066147dd3f service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Refreshing instance network info cache due to event network-changed-03b6ccf4-e23e-472e-a02a-e72bd1686c90. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1094.289958] env[61985]: DEBUG oslo_concurrency.lockutils [req-d64ce9fc-d75c-42f7-b323-5d3f975b6f10 req-8aeb0800-44a4-460a-b15c-6a066147dd3f service nova] Acquiring lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.290183] env[61985]: DEBUG oslo_concurrency.lockutils [req-d64ce9fc-d75c-42f7-b323-5d3f975b6f10 req-8aeb0800-44a4-460a-b15c-6a066147dd3f service nova] Acquired lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.290423] env[61985]: DEBUG nova.network.neutron [req-d64ce9fc-d75c-42f7-b323-5d3f975b6f10 req-8aeb0800-44a4-460a-b15c-6a066147dd3f service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Refreshing network info cache for port 03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1094.307274] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b2d1ec-8b60-d394-677c-b4408fed42ae, 'name': SearchDatastore_Task, 'duration_secs': 0.009595} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.307661] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.307975] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 926a2b59-8614-4e71-9d94-fd9feefdef46/926a2b59-8614-4e71-9d94-fd9feefdef46.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1094.308315] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ba8937a-7375-4da6-804f-746f9b04bfa7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.317969] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1094.317969] env[61985]: value = "task-936514" [ 1094.317969] env[61985]: _type = "Task" [ 1094.317969] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.326528] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.327943] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.328052] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.328155] env[61985]: DEBUG nova.network.neutron [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1094.414659] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936512, 'name': CreateVM_Task, 'duration_secs': 0.293623} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.414889] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1094.415679] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.415867] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.416260] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1094.416553] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54751b95-b470-441c-a283-946aa00bd242 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.421287] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1094.421287] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f8f59e-93ff-77c4-2737-1c26cedb566e" [ 1094.421287] env[61985]: _type = "Task" [ 1094.421287] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.433626] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f8f59e-93ff-77c4-2737-1c26cedb566e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.453418] env[61985]: DEBUG oslo_concurrency.lockutils [None req-515bb510-80ff-4289-a89c-0cc0c13adb78 tempest-AttachInterfacesTestJSON-1558490672 tempest-AttachInterfacesTestJSON-1558490672-project-member] Lock "5f2aa808-8d1e-471e-9ef7-0f91590ec546" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.275s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.633513] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6cb09e-f28d-4507-979f-9cde451e4ce8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.641178] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fee57b-211c-4943-a1e9-edb49bd1cfaf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.679211] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3868f1a5-7131-4939-963b-c9c1e1140605 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.687922] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be64bd3a-2e0d-4988-8886-5a3c991cdf3d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.704320] env[61985]: DEBUG nova.compute.provider_tree [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.752127] env[61985]: DEBUG oslo_vmware.api [None req-d662ccbb-89ed-40ef-b61d-ea63c5baeefb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936513, 'name': SuspendVM_Task} progress is 70%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.838025] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936514, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492399} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.838025] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 926a2b59-8614-4e71-9d94-fd9feefdef46/926a2b59-8614-4e71-9d94-fd9feefdef46.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1094.838025] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1094.838025] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e2086cb-6525-4b32-bc09-a06d32db1c52 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.848893] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1094.848893] env[61985]: value = "task-936515" [ 1094.848893] env[61985]: _type = "Task" [ 1094.848893] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.860616] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936515, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.932241] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f8f59e-93ff-77c4-2737-1c26cedb566e, 'name': SearchDatastore_Task, 'duration_secs': 0.018953} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.932534] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.932785] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1094.933053] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.933204] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.933390] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1094.933666] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85c9f8e3-7888-4def-a342-743646820d15 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.945130] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1094.945455] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1094.946275] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbc376e0-0138-4586-a57a-17d4bc34b477 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.951749] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1094.951749] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524c3eca-f86c-7586-2fba-a1e23a68b78b" [ 1094.951749] env[61985]: _type = "Task" [ 1094.951749] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.960640] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524c3eca-f86c-7586-2fba-a1e23a68b78b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.135209] env[61985]: DEBUG nova.network.neutron [req-d64ce9fc-d75c-42f7-b323-5d3f975b6f10 req-8aeb0800-44a4-460a-b15c-6a066147dd3f service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updated VIF entry in instance network info cache for port 03b6ccf4-e23e-472e-a02a-e72bd1686c90. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1095.136328] env[61985]: DEBUG nova.network.neutron [req-d64ce9fc-d75c-42f7-b323-5d3f975b6f10 req-8aeb0800-44a4-460a-b15c-6a066147dd3f service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating instance_info_cache with network_info: [{"id": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "address": "fa:16:3e:7c:35:40", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03b6ccf4-e2", "ovs_interfaceid": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.191414] env[61985]: DEBUG nova.network.neutron [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance_info_cache with network_info: [{"id": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "address": "fa:16:3e:13:20:58", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35f2970-0b", "ovs_interfaceid": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.207329] env[61985]: DEBUG nova.scheduler.client.report [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.252763] env[61985]: DEBUG oslo_vmware.api [None req-d662ccbb-89ed-40ef-b61d-ea63c5baeefb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936513, 'name': SuspendVM_Task, 'duration_secs': 0.600796} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.252763] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d662ccbb-89ed-40ef-b61d-ea63c5baeefb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Suspended the VM {{(pid=61985) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1095.252763] env[61985]: DEBUG nova.compute.manager [None req-d662ccbb-89ed-40ef-b61d-ea63c5baeefb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.252966] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90dff090-942c-46a5-9ed9-6af35b51be65 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.359672] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936515, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062334} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.359976] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1095.360746] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11893a74-ad00-4f9c-97f4-e6cef8a9e375 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.388735] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 926a2b59-8614-4e71-9d94-fd9feefdef46/926a2b59-8614-4e71-9d94-fd9feefdef46.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1095.390566] env[61985]: DEBUG nova.network.neutron [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updating instance_info_cache with network_info: [{"id": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "address": "fa:16:3e:af:c8:fd", "network": {"id": "d2cc00c1-7429-44f7-991c-ab44058f914a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1698886589-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be4f1b894ce04f37a0d106e93f4101fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47c6b710-95", "ovs_interfaceid": "47c6b710-9517-4551-b70b-d11e6cdd7c8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.395018] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6caa539d-d2a0-4b94-82fb-ea8b2742780b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.410523] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Releasing lock "refresh_cache-87006fbf-ea90-4d9a-88af-001de424ac14" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.411547] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb9f29d-3894-4658-bf1e-acd367a8bd72 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.418482] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Resuming the VM {{(pid=61985) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1095.418814] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86b84aec-2de6-4edb-ba81-0b7c2d938779 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.421581] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1095.421581] env[61985]: value = "task-936516" [ 1095.421581] env[61985]: _type = "Task" [ 1095.421581] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.426707] env[61985]: DEBUG oslo_vmware.api [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1095.426707] env[61985]: value = "task-936517" [ 1095.426707] env[61985]: _type = "Task" [ 1095.426707] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.435041] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936516, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.439539] env[61985]: DEBUG oslo_vmware.api [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936517, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.461524] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524c3eca-f86c-7586-2fba-a1e23a68b78b, 'name': SearchDatastore_Task, 'duration_secs': 0.008724} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.462354] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99c39476-3fda-42a2-a4e6-7d66b003bd60 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.467522] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1095.467522] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520a0a37-bf7c-9803-8de1-3e804d3f00f3" [ 1095.467522] env[61985]: _type = "Task" [ 1095.467522] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.475731] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520a0a37-bf7c-9803-8de1-3e804d3f00f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.509877] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0834be09-9edd-46fa-a690-9c4b959260e3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.529484] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance '885a36d9-24c6-407a-9d6b-01cd97e90674' progress to 0 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1095.639794] env[61985]: DEBUG oslo_concurrency.lockutils [req-d64ce9fc-d75c-42f7-b323-5d3f975b6f10 req-8aeb0800-44a4-460a-b15c-6a066147dd3f service nova] Releasing lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.694423] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.697461] env[61985]: DEBUG nova.compute.manager [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.698617] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe542988-8715-4ddd-a343-88cda77d20f0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.713093] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.714984] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.130s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.716384] env[61985]: INFO nova.compute.claims [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1095.741050] env[61985]: INFO nova.scheduler.client.report [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted allocations for instance 5aedbb9e-8bd1-499a-8195-b3045f7a141f [ 1095.935615] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936516, 'name': ReconfigVM_Task, 'duration_secs': 0.295187} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.935957] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 926a2b59-8614-4e71-9d94-fd9feefdef46/926a2b59-8614-4e71-9d94-fd9feefdef46.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1095.937548] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-692a7962-f54e-494e-808c-c3118a75e28d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.942654] env[61985]: DEBUG oslo_vmware.api [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936517, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.949310] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1095.949310] env[61985]: value = "task-936518" [ 1095.949310] env[61985]: _type = "Task" [ 1095.949310] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.960563] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936518, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.983046] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]520a0a37-bf7c-9803-8de1-3e804d3f00f3, 'name': SearchDatastore_Task, 'duration_secs': 0.009242} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.983360] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.983644] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1/0fac6d62-46b4-49a0-ab9a-f654ef27d1f1.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1095.983923] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9cd66e6-9233-44b7-b889-c2f18025c75d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.992557] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1095.992557] env[61985]: value = "task-936519" [ 1095.992557] env[61985]: _type = "Task" [ 1095.992557] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.002532] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936519, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.038640] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1096.038995] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ded968e5-49bc-4e9b-a69a-1f7f2940c78f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.047063] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1096.047063] env[61985]: value = "task-936520" [ 1096.047063] env[61985]: _type = "Task" [ 1096.047063] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.056090] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.250392] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3d9265b5-ed1d-490b-81b4-dd70cf142ad0 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "5aedbb9e-8bd1-499a-8195-b3045f7a141f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.011s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.438333] env[61985]: DEBUG oslo_vmware.api [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936517, 'name': PowerOnVM_Task, 'duration_secs': 0.520824} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.438979] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Resumed the VM {{(pid=61985) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1096.439305] env[61985]: DEBUG nova.compute.manager [None req-6956c5dd-a59f-4680-99e4-2a71eacaa129 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1096.441098] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b68d3a32-055c-4366-8b69-24493e965d04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.457931] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936518, 'name': Rename_Task, 'duration_secs': 0.161943} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.458492] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1096.459498] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18b50f73-bfa0-4b2c-97db-6921d1f63ebe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.466716] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1096.466716] env[61985]: value = "task-936521" [ 1096.466716] env[61985]: _type = "Task" [ 1096.466716] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.477653] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936521, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.502792] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936519, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453407} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.503807] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1/0fac6d62-46b4-49a0-ab9a-f654ef27d1f1.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1096.504288] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1096.504640] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-494293f1-d754-4acc-a5ba-4abdb01a8f39 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.511157] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1096.511157] env[61985]: value = "task-936522" [ 1096.511157] env[61985]: _type = "Task" [ 1096.511157] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.521042] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936522, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.558545] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.716255] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e376c8-1f26-4a4c-b9be-295d6eb0c8f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.736472] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Doing hard reboot of VM {{(pid=61985) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 1096.737874] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-4a12986a-c8af-4a5b-acfa-48e32a0bdf32 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.746072] env[61985]: DEBUG oslo_vmware.api [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1096.746072] env[61985]: value = "task-936523" [ 1096.746072] env[61985]: _type = "Task" [ 1096.746072] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.763024] env[61985]: DEBUG oslo_vmware.api [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936523, 'name': ResetVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.968490] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c098002-64dd-4e6e-bf03-2f261db3a6d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.985136] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936521, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.985793] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4baf800-8f5f-480f-add8-f00b484c5d0f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.022728] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ba7cda-4d4d-47d9-8efc-82518d5e2648 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.031033] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936522, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056672} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.033327] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1097.034152] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fde147b-f1fb-4be7-8ceb-4bcb0174eb3c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.037425] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99229472-7429-47b4-8713-2205524094b6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.051644] env[61985]: DEBUG nova.compute.provider_tree [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.078031] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1/0fac6d62-46b4-49a0-ab9a-f654ef27d1f1.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1097.079254] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9af562ec-aceb-409b-9f8b-81931d18f0a3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.100580] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.102069] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1097.102069] env[61985]: value = "task-936524" [ 1097.102069] env[61985]: _type = "Task" [ 1097.102069] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.111240] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936524, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.184333] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.184627] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.184845] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.185051] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.185239] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.187715] env[61985]: INFO nova.compute.manager [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Terminating instance [ 1097.189600] env[61985]: DEBUG nova.compute.manager [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1097.189826] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1097.190683] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2cd3cc-3c8b-47a2-bb63-30d24a074b83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.198795] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1097.199067] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-83b4b04b-f9e1-4cc0-a05f-f545eecbb11c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.255530] env[61985]: DEBUG oslo_vmware.api [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936523, 'name': ResetVM_Task, 'duration_secs': 0.09153} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.255804] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Did hard reboot of VM {{(pid=61985) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 1097.255984] env[61985]: DEBUG nova.compute.manager [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1097.257045] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3e66c5-0ca7-4ab9-bf39-29a1fabaff98 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.288161] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1097.289032] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1097.289282] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleting the datastore file [datastore1] 8cfc2489-b3b9-463d-9ad5-0cd73920ed58 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1097.289571] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a95d923c-bb73-4e45-8d74-510bcb8302a0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.296593] env[61985]: DEBUG oslo_vmware.api [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1097.296593] env[61985]: value = "task-936526" [ 1097.296593] env[61985]: _type = "Task" [ 1097.296593] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.309122] env[61985]: DEBUG oslo_vmware.api [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936526, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.484216] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936521, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.558058] env[61985]: DEBUG nova.scheduler.client.report [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1097.567461] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.611927] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936524, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.769178] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6328af54-557b-434c-b40a-b3474cda4c43 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.961s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.810074] env[61985]: DEBUG oslo_vmware.api [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936526, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.358604} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.810703] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1097.810999] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1097.811179] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1097.811407] env[61985]: INFO nova.compute.manager [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1097.811684] env[61985]: DEBUG oslo.service.loopingcall [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1097.811953] env[61985]: DEBUG nova.compute.manager [-] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1097.812111] env[61985]: DEBUG nova.network.neutron [-] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1097.987397] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936521, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.064666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.068191] env[61985]: DEBUG nova.compute.manager [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1098.068949] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.070340] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.334s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.070340] env[61985]: DEBUG nova.objects.instance [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lazy-loading 'resources' on Instance uuid acebe0a7-62d4-48d4-a1f7-5df0fa465753 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.093890] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "f45760fa-2298-422e-84ff-83f99b19627d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.094262] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "f45760fa-2298-422e-84ff-83f99b19627d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.113963] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936524, 'name': ReconfigVM_Task, 'duration_secs': 0.852721} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.114416] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1/0fac6d62-46b4-49a0-ab9a-f654ef27d1f1.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1098.115126] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d7d20372-0754-472e-bf7a-ed3d98eee19d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.122883] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1098.122883] env[61985]: value = "task-936527" [ 1098.122883] env[61985]: _type = "Task" [ 1098.122883] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.134305] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936527, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.276738] env[61985]: DEBUG nova.compute.manager [req-d5ac4e1d-c402-4a5b-a012-fd2317a62d74 req-c6286c51-7c45-4a88-9391-879fa8e0297c service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Received event network-vif-deleted-e1f6f60d-06ee-469d-bf35-a3af36741d55 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1098.276977] env[61985]: INFO nova.compute.manager [req-d5ac4e1d-c402-4a5b-a012-fd2317a62d74 req-c6286c51-7c45-4a88-9391-879fa8e0297c service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Neutron deleted interface e1f6f60d-06ee-469d-bf35-a3af36741d55; detaching it from the instance and deleting it from the info cache [ 1098.277180] env[61985]: DEBUG nova.network.neutron [req-d5ac4e1d-c402-4a5b-a012-fd2317a62d74 req-c6286c51-7c45-4a88-9391-879fa8e0297c service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.497239] env[61985]: DEBUG oslo_vmware.api [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936521, 'name': PowerOnVM_Task, 'duration_secs': 1.549804} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.497239] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1098.497239] env[61985]: INFO nova.compute.manager [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Took 9.04 seconds to spawn the instance on the hypervisor. [ 1098.497239] env[61985]: DEBUG nova.compute.manager [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1098.497239] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea108871-a11f-4eda-82bd-d8c0f7b97dc2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.565063] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.570451] env[61985]: DEBUG nova.compute.utils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1098.571779] env[61985]: DEBUG nova.compute.manager [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1098.571954] env[61985]: DEBUG nova.network.neutron [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1098.597217] env[61985]: DEBUG nova.compute.manager [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1098.632258] env[61985]: DEBUG nova.policy [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc86b7a24ed94efbae2dc437e4fef389', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '60ecaef69401418eac6c655c1b7df5be', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1098.640459] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936527, 'name': Rename_Task, 'duration_secs': 0.264865} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.643817] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1098.644546] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-486912b2-0aa9-4e53-bf31-8df111d20f8c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.651966] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1098.651966] env[61985]: value = "task-936528" [ 1098.651966] env[61985]: _type = "Task" [ 1098.651966] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.666637] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936528, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.719912] env[61985]: DEBUG nova.network.neutron [-] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.780092] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a133db04-e0f6-409e-8854-c5e4edc8e28f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.792437] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3049bb81-aa43-4cb9-9cd9-afc9d1190748 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.837473] env[61985]: DEBUG nova.compute.manager [req-d5ac4e1d-c402-4a5b-a012-fd2317a62d74 req-c6286c51-7c45-4a88-9391-879fa8e0297c service nova] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Detach interface failed, port_id=e1f6f60d-06ee-469d-bf35-a3af36741d55, reason: Instance 8cfc2489-b3b9-463d-9ad5-0cd73920ed58 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1098.840139] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2060da9a-5c2e-4ba5-8efd-de2c8dd41e2f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.849617] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42793c4-0a87-44ed-810b-27c923de9135 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.901130] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715e25a1-b826-47b1-9c54-d0b90d94048e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.911955] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c058cc-deb3-484a-baa4-23dd7eafb024 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.932328] env[61985]: DEBUG nova.compute.provider_tree [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.018558] env[61985]: INFO nova.compute.manager [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Took 19.15 seconds to build instance. [ 1099.066875] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936520, 'name': PowerOffVM_Task, 'duration_secs': 2.699672} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.067192] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1099.067387] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance '885a36d9-24c6-407a-9d6b-01cd97e90674' progress to 17 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1099.076831] env[61985]: DEBUG nova.compute.manager [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1099.124021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.164430] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936528, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.169025] env[61985]: DEBUG nova.network.neutron [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Successfully created port: 0aa5789a-a605-4e56-aa4e-0402756b8569 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1099.223084] env[61985]: INFO nova.compute.manager [-] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Took 1.41 seconds to deallocate network for instance. [ 1099.435314] env[61985]: DEBUG nova.scheduler.client.report [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.520918] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a3cb66a-d9be-489d-8b8a-0d2baf19d5bb tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "926a2b59-8614-4e71-9d94-fd9feefdef46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.678s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.574311] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1099.574590] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1099.574731] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1099.574920] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1099.575087] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1099.575246] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1099.575455] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1099.575623] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1099.575799] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1099.576534] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1099.576777] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1099.588099] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce0ac86e-f650-44e6-a381-5153440f618c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.605359] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1099.605359] env[61985]: value = "task-936529" [ 1099.605359] env[61985]: _type = "Task" [ 1099.605359] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.614582] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936529, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.663773] env[61985]: DEBUG oslo_vmware.api [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936528, 'name': PowerOnVM_Task, 'duration_secs': 0.800151} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.664297] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1099.664516] env[61985]: INFO nova.compute.manager [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Took 7.87 seconds to spawn the instance on the hypervisor. [ 1099.664831] env[61985]: DEBUG nova.compute.manager [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1099.665920] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7756092-dbc9-4bf4-b5c2-0ce9c688d52c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.730494] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.803706] env[61985]: DEBUG nova.network.neutron [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Successfully created port: 188cbfa9-832a-4537-ba42-d51325778b56 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1099.940805] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.871s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.947031] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.823s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.948984] env[61985]: INFO nova.compute.claims [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1099.980240] env[61985]: INFO nova.scheduler.client.report [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Deleted allocations for instance acebe0a7-62d4-48d4-a1f7-5df0fa465753 [ 1100.100270] env[61985]: DEBUG nova.compute.manager [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1100.115551] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936529, 'name': ReconfigVM_Task, 'duration_secs': 0.296499} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.115899] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance '885a36d9-24c6-407a-9d6b-01cd97e90674' progress to 33 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1100.139156] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1100.139674] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1100.139674] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1100.139855] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1100.140469] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1100.140704] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1100.141096] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1100.141268] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1100.141687] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1100.141941] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1100.142129] env[61985]: DEBUG nova.virt.hardware [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1100.143285] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219c0409-de8e-4b3b-b607-8da02a245ede {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.154570] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c985330-569a-4f70-a59a-54b65a736765 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.187904] env[61985]: INFO nova.compute.manager [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Took 18.34 seconds to build instance. [ 1100.498043] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f3bf234-5c75-4533-aad4-823f6c2b1d0c tempest-ServersNegativeTestMultiTenantJSON-352614885 tempest-ServersNegativeTestMultiTenantJSON-352614885-project-member] Lock "acebe0a7-62d4-48d4-a1f7-5df0fa465753" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.804s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.624867] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1100.626023] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1100.626023] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1100.626023] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1100.626023] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1100.626023] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1100.626713] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1100.628819] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1100.629144] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1100.629546] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1100.629803] env[61985]: DEBUG nova.virt.hardware [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1100.639655] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfiguring VM instance instance-0000005a to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1100.639999] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1f92172-c58e-42cf-acdc-48d5652a3aea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.661847] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1100.661847] env[61985]: value = "task-936530" [ 1100.661847] env[61985]: _type = "Task" [ 1100.661847] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.671700] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936530, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.693276] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bb5c4b78-fec8-42ae-aeb6-598734eb987f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.858s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.180356] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936530, 'name': ReconfigVM_Task, 'duration_secs': 0.198391} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.180356] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfigured VM instance instance-0000005a to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1101.180833] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03620c6a-bab9-401e-b8a9-c0b14843bd70 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.184742] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8188953e-c340-4a7f-b4a1-15feb3b41cca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.207933] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc2ddf5-1a7e-4b17-87ce-e560ce1d334e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.219692] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 885a36d9-24c6-407a-9d6b-01cd97e90674/885a36d9-24c6-407a-9d6b-01cd97e90674.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.219692] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65685d97-d67d-4a49-a8f8-78b3fd2577ae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.265807] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39382f9-8003-4230-ba25-4daca7b8b61e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.268654] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1101.268654] env[61985]: value = "task-936531" [ 1101.268654] env[61985]: _type = "Task" [ 1101.268654] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.275476] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3de6aaf-4c57-4cfa-8103-eff11307fd2e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.282659] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.292776] env[61985]: DEBUG nova.compute.provider_tree [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.785124] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936531, 'name': ReconfigVM_Task, 'duration_secs': 0.406148} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.786080] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 885a36d9-24c6-407a-9d6b-01cd97e90674/885a36d9-24c6-407a-9d6b-01cd97e90674.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1101.787356] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance '885a36d9-24c6-407a-9d6b-01cd97e90674' progress to 50 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1101.795115] env[61985]: DEBUG nova.scheduler.client.report [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1102.102402] env[61985]: DEBUG nova.compute.manager [req-eadf5857-3f83-4f13-9458-a3a7bd212ca7 req-77b1cee6-8e34-4459-b2e1-7938724f7bbd service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Received event network-vif-plugged-0aa5789a-a605-4e56-aa4e-0402756b8569 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1102.102598] env[61985]: DEBUG oslo_concurrency.lockutils [req-eadf5857-3f83-4f13-9458-a3a7bd212ca7 req-77b1cee6-8e34-4459-b2e1-7938724f7bbd service nova] Acquiring lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.102836] env[61985]: DEBUG oslo_concurrency.lockutils [req-eadf5857-3f83-4f13-9458-a3a7bd212ca7 req-77b1cee6-8e34-4459-b2e1-7938724f7bbd service nova] Lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.103017] env[61985]: DEBUG oslo_concurrency.lockutils [req-eadf5857-3f83-4f13-9458-a3a7bd212ca7 req-77b1cee6-8e34-4459-b2e1-7938724f7bbd service nova] Lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.103365] env[61985]: DEBUG nova.compute.manager [req-eadf5857-3f83-4f13-9458-a3a7bd212ca7 req-77b1cee6-8e34-4459-b2e1-7938724f7bbd service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] No waiting events found dispatching network-vif-plugged-0aa5789a-a605-4e56-aa4e-0402756b8569 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1102.103953] env[61985]: WARNING nova.compute.manager [req-eadf5857-3f83-4f13-9458-a3a7bd212ca7 req-77b1cee6-8e34-4459-b2e1-7938724f7bbd service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Received unexpected event network-vif-plugged-0aa5789a-a605-4e56-aa4e-0402756b8569 for instance with vm_state building and task_state spawning. [ 1102.154030] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "926a2b59-8614-4e71-9d94-fd9feefdef46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.154030] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "926a2b59-8614-4e71-9d94-fd9feefdef46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.154030] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "926a2b59-8614-4e71-9d94-fd9feefdef46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.154030] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "926a2b59-8614-4e71-9d94-fd9feefdef46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.154030] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "926a2b59-8614-4e71-9d94-fd9feefdef46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.157508] env[61985]: INFO nova.compute.manager [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Terminating instance [ 1102.160276] env[61985]: DEBUG nova.compute.manager [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1102.160509] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1102.161553] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162cfb33-0726-466a-8162-1ee7aff75fd0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.173395] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1102.174736] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7238dfc-cc95-4738-9163-efe92ed22693 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.185821] env[61985]: DEBUG oslo_vmware.api [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1102.185821] env[61985]: value = "task-936532" [ 1102.185821] env[61985]: _type = "Task" [ 1102.185821] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.200813] env[61985]: DEBUG oslo_vmware.api [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936532, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.234126] env[61985]: DEBUG nova.network.neutron [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Successfully updated port: 0aa5789a-a605-4e56-aa4e-0402756b8569 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1102.235887] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquiring lock "ef330828-c218-4b80-8b1a-6813bdb67c4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.236615] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lock "ef330828-c218-4b80-8b1a-6813bdb67c4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.295874] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd7ec07-58e4-423a-9821-e73f86e00985 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.300447] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.303018] env[61985]: DEBUG nova.compute.manager [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1102.304436] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.574s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.304968] env[61985]: DEBUG nova.objects.instance [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lazy-loading 'resources' on Instance uuid 8cfc2489-b3b9-463d-9ad5-0cd73920ed58 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.332994] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f209dd-2505-4b9c-a190-66dd3afb9325 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.357216] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance '885a36d9-24c6-407a-9d6b-01cd97e90674' progress to 67 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1102.695283] env[61985]: DEBUG oslo_vmware.api [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936532, 'name': PowerOffVM_Task, 'duration_secs': 0.198804} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.695626] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1102.695840] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1102.696152] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a0f5ecd2-d5bf-4f3c-87d5-50f4cae1d4f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.741457] env[61985]: DEBUG nova.compute.manager [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1102.808289] env[61985]: DEBUG nova.compute.utils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1102.809999] env[61985]: DEBUG nova.compute.manager [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1102.814434] env[61985]: DEBUG nova.network.neutron [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1102.850045] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1102.850045] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1102.850406] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleting the datastore file [datastore2] 926a2b59-8614-4e71-9d94-fd9feefdef46 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.850880] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb0b8f49-3095-4145-93fe-4e4dd7123e2f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.861102] env[61985]: DEBUG oslo_vmware.api [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for the task: (returnval){ [ 1102.861102] env[61985]: value = "task-936534" [ 1102.861102] env[61985]: _type = "Task" [ 1102.861102] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.876840] env[61985]: DEBUG oslo_vmware.api [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936534, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.911737] env[61985]: DEBUG nova.policy [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e23fdb33638143409613d9d4f26f199c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe531ec9f60548adb457351e19f8d51b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1103.051024] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea440d2-631f-4783-afc0-0733c76c7c93 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.056710] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195be1cf-54c6-4a2a-b79d-1ad88767c232 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.090887] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5794ba50-6092-468e-a312-b405ef25c0f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.103033] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82853684-9668-4979-81b6-047c7e8dcde5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.123223] env[61985]: DEBUG nova.compute.provider_tree [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.264230] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.318797] env[61985]: DEBUG nova.compute.manager [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1103.371459] env[61985]: DEBUG nova.network.neutron [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Successfully created port: 650759b1-a52a-4181-ac42-6744da8673c9 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1103.381872] env[61985]: DEBUG oslo_vmware.api [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Task: {'id': task-936534, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.409478} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.382227] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.382536] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1103.382845] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1103.383168] env[61985]: INFO nova.compute.manager [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Took 1.22 seconds to destroy the instance on the hypervisor. [ 1103.383553] env[61985]: DEBUG oslo.service.loopingcall [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.386471] env[61985]: DEBUG nova.compute.manager [-] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1103.386591] env[61985]: DEBUG nova.network.neutron [-] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1103.613969] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "87006fbf-ea90-4d9a-88af-001de424ac14" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.614306] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "87006fbf-ea90-4d9a-88af-001de424ac14" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.614522] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.614716] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.614894] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "87006fbf-ea90-4d9a-88af-001de424ac14-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.617106] env[61985]: INFO nova.compute.manager [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Terminating instance [ 1103.618970] env[61985]: DEBUG nova.compute.manager [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1103.619178] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1103.620008] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aff2bb3-3220-4e79-b96c-1dcbe1f2b459 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.627936] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1103.628173] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4bdb840e-dcda-40e0-9c1c-6f593a3e9d9b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.632218] env[61985]: DEBUG nova.scheduler.client.report [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1103.636545] env[61985]: DEBUG oslo_vmware.api [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1103.636545] env[61985]: value = "task-936535" [ 1103.636545] env[61985]: _type = "Task" [ 1103.636545] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.656092] env[61985]: DEBUG oslo_vmware.api [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936535, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.046763] env[61985]: DEBUG nova.compute.manager [req-9ce06b4c-b251-4f21-bd41-8c11e89663d6 req-d3f3f82a-2bfb-4896-90b2-10ddd544aab4 service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Received event network-vif-deleted-4249c70c-abfc-4d75-b161-1bfda2047dd6 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1104.047009] env[61985]: INFO nova.compute.manager [req-9ce06b4c-b251-4f21-bd41-8c11e89663d6 req-d3f3f82a-2bfb-4896-90b2-10ddd544aab4 service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Neutron deleted interface 4249c70c-abfc-4d75-b161-1bfda2047dd6; detaching it from the instance and deleting it from the info cache [ 1104.047376] env[61985]: DEBUG nova.network.neutron [req-9ce06b4c-b251-4f21-bd41-8c11e89663d6 req-d3f3f82a-2bfb-4896-90b2-10ddd544aab4 service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.079093] env[61985]: DEBUG nova.network.neutron [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Port a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1104.142137] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.833s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.142137] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.877s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.143434] env[61985]: INFO nova.compute.claims [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1104.150782] env[61985]: DEBUG nova.compute.manager [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received event network-changed-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1104.150964] env[61985]: DEBUG nova.compute.manager [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Refreshing instance network info cache due to event network-changed-03b6ccf4-e23e-472e-a02a-e72bd1686c90. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1104.151200] env[61985]: DEBUG oslo_concurrency.lockutils [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] Acquiring lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.151594] env[61985]: DEBUG oslo_concurrency.lockutils [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] Acquired lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.151594] env[61985]: DEBUG nova.network.neutron [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Refreshing network info cache for port 03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1104.159147] env[61985]: DEBUG oslo_vmware.api [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936535, 'name': PowerOffVM_Task, 'duration_secs': 0.208625} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.159624] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1104.159791] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1104.160268] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-398bf713-6a54-41a7-8b35-57557bf1372c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.180224] env[61985]: INFO nova.scheduler.client.report [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted allocations for instance 8cfc2489-b3b9-463d-9ad5-0cd73920ed58 [ 1104.330120] env[61985]: DEBUG nova.compute.manager [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1104.358080] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1104.358212] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1104.358373] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1104.358527] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1104.358765] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1104.358958] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1104.359343] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1104.359567] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1104.359709] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1104.359882] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1104.360151] env[61985]: DEBUG nova.virt.hardware [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1104.361056] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f46775e-6ad9-428d-9d6f-c2512ecc2d7b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.369592] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e578b16a-82fc-459d-80f0-66ff3db9830d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.408791] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1104.408791] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1104.408791] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleting the datastore file [datastore2] 87006fbf-ea90-4d9a-88af-001de424ac14 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1104.409073] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac3ae4fc-95d0-4066-8e7e-ff8528c5fbf8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.416614] env[61985]: DEBUG oslo_vmware.api [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for the task: (returnval){ [ 1104.416614] env[61985]: value = "task-936537" [ 1104.416614] env[61985]: _type = "Task" [ 1104.416614] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.427591] env[61985]: DEBUG oslo_vmware.api [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936537, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.463325] env[61985]: DEBUG nova.network.neutron [-] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.486021] env[61985]: DEBUG nova.network.neutron [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Successfully updated port: 188cbfa9-832a-4537-ba42-d51325778b56 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1104.552735] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7537e66f-8ce6-451e-94de-aff5d1092f5b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.569301] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e323344-cb9b-4777-97de-e69c9fee2a48 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.599481] env[61985]: DEBUG nova.compute.manager [req-9ce06b4c-b251-4f21-bd41-8c11e89663d6 req-d3f3f82a-2bfb-4896-90b2-10ddd544aab4 service nova] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Detach interface failed, port_id=4249c70c-abfc-4d75-b161-1bfda2047dd6, reason: Instance 926a2b59-8614-4e71-9d94-fd9feefdef46 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1104.694202] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa8da7e6-056c-4dfa-b9bf-1b588e6e3bbb tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "8cfc2489-b3b9-463d-9ad5-0cd73920ed58" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.509s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.926212] env[61985]: DEBUG oslo_vmware.api [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Task: {'id': task-936537, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147423} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.926558] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1104.926756] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1104.926941] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1104.927147] env[61985]: INFO nova.compute.manager [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Took 1.31 seconds to destroy the instance on the hypervisor. [ 1104.927409] env[61985]: DEBUG oslo.service.loopingcall [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1104.927613] env[61985]: DEBUG nova.compute.manager [-] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1104.927727] env[61985]: DEBUG nova.network.neutron [-] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1104.934346] env[61985]: DEBUG nova.network.neutron [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updated VIF entry in instance network info cache for port 03b6ccf4-e23e-472e-a02a-e72bd1686c90. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1104.934770] env[61985]: DEBUG nova.network.neutron [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating instance_info_cache with network_info: [{"id": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "address": "fa:16:3e:7c:35:40", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03b6ccf4-e2", "ovs_interfaceid": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.966266] env[61985]: INFO nova.compute.manager [-] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Took 1.58 seconds to deallocate network for instance. [ 1104.985924] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "refresh_cache-93dd0394-d7e1-42d5-a58f-fcd621c044b3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.986673] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquired lock "refresh_cache-93dd0394-d7e1-42d5-a58f-fcd621c044b3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.987963] env[61985]: DEBUG nova.network.neutron [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1105.104012] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "885a36d9-24c6-407a-9d6b-01cd97e90674-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.104332] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.104552] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.368640] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1c7592-3329-49b2-9bc3-10e4266a6571 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.376690] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c36c80c-d769-4eef-90e2-5b5aac9a8636 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.411328] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15bc6053-36d5-49c7-b750-766d1309803d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.420392] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c23abc9-b94f-45fd-8c32-6d2678ca8f86 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.436230] env[61985]: DEBUG nova.compute.provider_tree [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1105.437761] env[61985]: DEBUG oslo_concurrency.lockutils [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] Releasing lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.438034] env[61985]: DEBUG nova.compute.manager [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Received event network-changed-0aa5789a-a605-4e56-aa4e-0402756b8569 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1105.438221] env[61985]: DEBUG nova.compute.manager [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Refreshing instance network info cache due to event network-changed-0aa5789a-a605-4e56-aa4e-0402756b8569. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1105.438409] env[61985]: DEBUG oslo_concurrency.lockutils [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] Acquiring lock "refresh_cache-93dd0394-d7e1-42d5-a58f-fcd621c044b3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.473210] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.483618] env[61985]: DEBUG nova.network.neutron [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Successfully updated port: 650759b1-a52a-4181-ac42-6744da8673c9 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1105.526436] env[61985]: DEBUG nova.network.neutron [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1105.769408] env[61985]: DEBUG nova.network.neutron [-] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.837208] env[61985]: DEBUG nova.network.neutron [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Updating instance_info_cache with network_info: [{"id": "0aa5789a-a605-4e56-aa4e-0402756b8569", "address": "fa:16:3e:fa:b3:fa", "network": {"id": "1f73c5a9-e080-48bf-922d-8d4739af644c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-367732363", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0aa5789a-a6", "ovs_interfaceid": "0aa5789a-a605-4e56-aa4e-0402756b8569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "188cbfa9-832a-4537-ba42-d51325778b56", "address": "fa:16:3e:48:62:16", "network": {"id": "21e4b9d5-4291-4813-bbc7-020d922cff31", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1313355272", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.80", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap188cbfa9-83", "ovs_interfaceid": "188cbfa9-832a-4537-ba42-d51325778b56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.940245] env[61985]: DEBUG nova.scheduler.client.report [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1105.973449] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "036fae78-2af0-49f1-85fa-245c29f35ebc" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.974100] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.986234] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "refresh_cache-f45760fa-2298-422e-84ff-83f99b19627d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.986234] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "refresh_cache-f45760fa-2298-422e-84ff-83f99b19627d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.986380] env[61985]: DEBUG nova.network.neutron [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1106.076721] env[61985]: DEBUG nova.compute.manager [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Received event network-vif-deleted-47c6b710-9517-4551-b70b-d11e6cdd7c8f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1106.076940] env[61985]: DEBUG nova.compute.manager [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Received event network-vif-plugged-650759b1-a52a-4181-ac42-6744da8673c9 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1106.077160] env[61985]: DEBUG oslo_concurrency.lockutils [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] Acquiring lock "f45760fa-2298-422e-84ff-83f99b19627d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.077390] env[61985]: DEBUG oslo_concurrency.lockutils [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] Lock "f45760fa-2298-422e-84ff-83f99b19627d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.077586] env[61985]: DEBUG oslo_concurrency.lockutils [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] Lock "f45760fa-2298-422e-84ff-83f99b19627d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.077810] env[61985]: DEBUG nova.compute.manager [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] [instance: f45760fa-2298-422e-84ff-83f99b19627d] No waiting events found dispatching network-vif-plugged-650759b1-a52a-4181-ac42-6744da8673c9 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1106.078009] env[61985]: WARNING nova.compute.manager [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Received unexpected event network-vif-plugged-650759b1-a52a-4181-ac42-6744da8673c9 for instance with vm_state building and task_state spawning. [ 1106.078185] env[61985]: DEBUG nova.compute.manager [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Received event network-changed-650759b1-a52a-4181-ac42-6744da8673c9 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1106.078364] env[61985]: DEBUG nova.compute.manager [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Refreshing instance network info cache due to event network-changed-650759b1-a52a-4181-ac42-6744da8673c9. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1106.078539] env[61985]: DEBUG oslo_concurrency.lockutils [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] Acquiring lock "refresh_cache-f45760fa-2298-422e-84ff-83f99b19627d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.108326] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.108578] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.152350] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.152544] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.152724] env[61985]: DEBUG nova.network.neutron [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1106.174478] env[61985]: DEBUG nova.compute.manager [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Received event network-vif-plugged-188cbfa9-832a-4537-ba42-d51325778b56 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1106.174618] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] Acquiring lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.174787] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] Lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.174942] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] Lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.175132] env[61985]: DEBUG nova.compute.manager [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] No waiting events found dispatching network-vif-plugged-188cbfa9-832a-4537-ba42-d51325778b56 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1106.176026] env[61985]: WARNING nova.compute.manager [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Received unexpected event network-vif-plugged-188cbfa9-832a-4537-ba42-d51325778b56 for instance with vm_state building and task_state spawning. [ 1106.176026] env[61985]: DEBUG nova.compute.manager [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Received event network-changed-188cbfa9-832a-4537-ba42-d51325778b56 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1106.176026] env[61985]: DEBUG nova.compute.manager [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Refreshing instance network info cache due to event network-changed-188cbfa9-832a-4537-ba42-d51325778b56. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1106.176026] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] Acquiring lock "refresh_cache-93dd0394-d7e1-42d5-a58f-fcd621c044b3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.273260] env[61985]: INFO nova.compute.manager [-] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Took 1.34 seconds to deallocate network for instance. [ 1106.340649] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Releasing lock "refresh_cache-93dd0394-d7e1-42d5-a58f-fcd621c044b3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.341172] env[61985]: DEBUG nova.compute.manager [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Instance network_info: |[{"id": "0aa5789a-a605-4e56-aa4e-0402756b8569", "address": "fa:16:3e:fa:b3:fa", "network": {"id": "1f73c5a9-e080-48bf-922d-8d4739af644c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-367732363", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0aa5789a-a6", "ovs_interfaceid": "0aa5789a-a605-4e56-aa4e-0402756b8569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "188cbfa9-832a-4537-ba42-d51325778b56", "address": "fa:16:3e:48:62:16", "network": {"id": "21e4b9d5-4291-4813-bbc7-020d922cff31", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1313355272", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.80", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap188cbfa9-83", "ovs_interfaceid": "188cbfa9-832a-4537-ba42-d51325778b56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1106.341541] env[61985]: DEBUG oslo_concurrency.lockutils [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] Acquired lock "refresh_cache-93dd0394-d7e1-42d5-a58f-fcd621c044b3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.341780] env[61985]: DEBUG nova.network.neutron [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Refreshing network info cache for port 0aa5789a-a605-4e56-aa4e-0402756b8569 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1106.342971] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:b3:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20fbc86c-0935-406c-9659-3ffc33fc0d08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0aa5789a-a605-4e56-aa4e-0402756b8569', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:62:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '504a3211-8288-4b87-a52a-590466d80a49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '188cbfa9-832a-4537-ba42-d51325778b56', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1106.352437] env[61985]: DEBUG oslo.service.loopingcall [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1106.353337] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1106.353578] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-611c8586-0f48-4834-a8be-f39097343328 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.376071] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1106.376071] env[61985]: value = "task-936538" [ 1106.376071] env[61985]: _type = "Task" [ 1106.376071] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.383807] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936538, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.445795] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.446532] env[61985]: DEBUG nova.compute.manager [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1106.449503] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.976s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.449738] env[61985]: DEBUG nova.objects.instance [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lazy-loading 'resources' on Instance uuid 926a2b59-8614-4e71-9d94-fd9feefdef46 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.476938] env[61985]: DEBUG nova.compute.utils [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1106.515334] env[61985]: DEBUG nova.network.neutron [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1106.613680] env[61985]: DEBUG nova.compute.manager [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1106.661841] env[61985]: DEBUG nova.network.neutron [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Updating instance_info_cache with network_info: [{"id": "650759b1-a52a-4181-ac42-6744da8673c9", "address": "fa:16:3e:4d:d6:97", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap650759b1-a5", "ovs_interfaceid": "650759b1-a52a-4181-ac42-6744da8673c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.779788] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.886786] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936538, 'name': CreateVM_Task, 'duration_secs': 0.405406} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.886963] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1106.891019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.891019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.891019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1106.891019] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c881fa36-1a80-45e4-92db-bee5e536f6b7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.894641] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1106.894641] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525cd492-e536-a5b8-a0d7-c5056a24736f" [ 1106.894641] env[61985]: _type = "Task" [ 1106.894641] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.902966] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525cd492-e536-a5b8-a0d7-c5056a24736f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.924318] env[61985]: DEBUG nova.network.neutron [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance_info_cache with network_info: [{"id": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "address": "fa:16:3e:e1:d0:1f", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64c525f-b3", "ovs_interfaceid": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.954760] env[61985]: DEBUG nova.compute.utils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1106.959058] env[61985]: DEBUG nova.compute.manager [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1106.959058] env[61985]: DEBUG nova.network.neutron [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1106.979963] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.028189] env[61985]: DEBUG nova.policy [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c9df7a9f8f8e464eb21fd284e1968030', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fa7a4522375426f9f3cc76e7f9270a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1107.146634] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.166547] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "refresh_cache-f45760fa-2298-422e-84ff-83f99b19627d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.166547] env[61985]: DEBUG nova.compute.manager [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Instance network_info: |[{"id": "650759b1-a52a-4181-ac42-6744da8673c9", "address": "fa:16:3e:4d:d6:97", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap650759b1-a5", "ovs_interfaceid": "650759b1-a52a-4181-ac42-6744da8673c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1107.166547] env[61985]: DEBUG oslo_concurrency.lockutils [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] Acquired lock "refresh_cache-f45760fa-2298-422e-84ff-83f99b19627d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.166547] env[61985]: DEBUG nova.network.neutron [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Refreshing network info cache for port 650759b1-a52a-4181-ac42-6744da8673c9 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1107.167419] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:d6:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55520f67-d092-4eb7-940f-d7cceaa1ca1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '650759b1-a52a-4181-ac42-6744da8673c9', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1107.176091] env[61985]: DEBUG oslo.service.loopingcall [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1107.176091] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1107.176091] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed2f6061-91bc-4a2b-850e-a4bd5de8b6c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.199967] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1107.199967] env[61985]: value = "task-936539" [ 1107.199967] env[61985]: _type = "Task" [ 1107.199967] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.215665] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936539, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.228218] env[61985]: DEBUG nova.network.neutron [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Updated VIF entry in instance network info cache for port 0aa5789a-a605-4e56-aa4e-0402756b8569. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1107.228218] env[61985]: DEBUG nova.network.neutron [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Updating instance_info_cache with network_info: [{"id": "0aa5789a-a605-4e56-aa4e-0402756b8569", "address": "fa:16:3e:fa:b3:fa", "network": {"id": "1f73c5a9-e080-48bf-922d-8d4739af644c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-367732363", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0aa5789a-a6", "ovs_interfaceid": "0aa5789a-a605-4e56-aa4e-0402756b8569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "188cbfa9-832a-4537-ba42-d51325778b56", "address": "fa:16:3e:48:62:16", "network": {"id": "21e4b9d5-4291-4813-bbc7-020d922cff31", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1313355272", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.80", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap188cbfa9-83", "ovs_interfaceid": "188cbfa9-832a-4537-ba42-d51325778b56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.248973] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf776a6-7a6b-4402-86dd-b2dcc12988b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.260015] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0549bd1-6b35-4015-8a45-0a3621547ac0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.294480] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebb8795-9d11-4f2d-8d68-2c4671dc0e84 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.302217] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a416cf6-fadb-428a-a8e3-dded5c9b579b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.316290] env[61985]: DEBUG nova.compute.provider_tree [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1107.400059] env[61985]: DEBUG nova.network.neutron [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Successfully created port: ecf04033-48b6-4b81-ab17-57a89dc0be60 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1107.408022] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525cd492-e536-a5b8-a0d7-c5056a24736f, 'name': SearchDatastore_Task, 'duration_secs': 0.009405} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.408377] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.408669] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1107.409036] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.409633] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.409633] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1107.409791] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cfd85fde-ae16-43c7-8bbd-84134552c965 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.418707] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1107.418906] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1107.419945] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0508135a-bf71-4b90-b72d-7b4f8036a94e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.425762] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1107.425762] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523963f1-7385-5e15-17b5-1b6fde66d84d" [ 1107.425762] env[61985]: _type = "Task" [ 1107.425762] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.429925] env[61985]: DEBUG oslo_concurrency.lockutils [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.438801] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523963f1-7385-5e15-17b5-1b6fde66d84d, 'name': SearchDatastore_Task, 'duration_secs': 0.009285} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.439576] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbd5347c-8969-4c16-81c6-e5b0300ab7ba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.444758] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1107.444758] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52092882-05fa-0220-02e9-fcfb57229cef" [ 1107.444758] env[61985]: _type = "Task" [ 1107.444758] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.456074] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52092882-05fa-0220-02e9-fcfb57229cef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.463038] env[61985]: DEBUG nova.compute.manager [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1107.711416] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936539, 'name': CreateVM_Task, 'duration_secs': 0.323551} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.711551] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1107.712239] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.712408] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.712745] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1107.713009] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8390dec-d0ef-4b23-8c68-90ec42d0159a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.717907] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1107.717907] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52dde393-63eb-f2a0-0640-9ce1a84a41e1" [ 1107.717907] env[61985]: _type = "Task" [ 1107.717907] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.725469] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52dde393-63eb-f2a0-0640-9ce1a84a41e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.730238] env[61985]: DEBUG oslo_concurrency.lockutils [req-991c59e8-28f7-4df8-b37e-e0f52abe03fe req-89bd09bf-4b42-422b-9e5b-d1235c41912e service nova] Releasing lock "refresh_cache-93dd0394-d7e1-42d5-a58f-fcd621c044b3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.730890] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] Acquired lock "refresh_cache-93dd0394-d7e1-42d5-a58f-fcd621c044b3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.730890] env[61985]: DEBUG nova.network.neutron [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Refreshing network info cache for port 188cbfa9-832a-4537-ba42-d51325778b56 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1107.819380] env[61985]: DEBUG nova.scheduler.client.report [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1107.888610] env[61985]: DEBUG nova.network.neutron [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Updated VIF entry in instance network info cache for port 650759b1-a52a-4181-ac42-6744da8673c9. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1107.889084] env[61985]: DEBUG nova.network.neutron [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Updating instance_info_cache with network_info: [{"id": "650759b1-a52a-4181-ac42-6744da8673c9", "address": "fa:16:3e:4d:d6:97", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap650759b1-a5", "ovs_interfaceid": "650759b1-a52a-4181-ac42-6744da8673c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.940492] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f06cf7-377d-44b4-95c1-2c4df1559321 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.950346] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4081612c-c4fa-46a0-b14f-2057753821d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.957397] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52092882-05fa-0220-02e9-fcfb57229cef, 'name': SearchDatastore_Task, 'duration_secs': 0.008942} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.957668] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.957929] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 93dd0394-d7e1-42d5-a58f-fcd621c044b3/93dd0394-d7e1-42d5-a58f-fcd621c044b3.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1107.958180] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb85f93f-b7b8-43db-ab27-5312cfcf54e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.966643] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1107.966643] env[61985]: value = "task-936540" [ 1107.966643] env[61985]: _type = "Task" [ 1107.966643] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.976596] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.078497] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "036fae78-2af0-49f1-85fa-245c29f35ebc" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.078688] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.078963] env[61985]: INFO nova.compute.manager [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Attaching volume 9d006806-6309-415f-b72a-5e327ec77248 to /dev/sdb [ 1108.110815] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b5f63a-ebbc-4670-a092-5e84e18a20cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.117548] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf53f536-1dfd-44ea-b437-5e5150a48899 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.131053] env[61985]: DEBUG nova.virt.block_device [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Updating existing volume attachment record: 04997e43-4838-4f52-8b1e-1dbc35034060 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1108.228253] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52dde393-63eb-f2a0-0640-9ce1a84a41e1, 'name': SearchDatastore_Task, 'duration_secs': 0.009995} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.228687] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.228887] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1108.229188] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.229472] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.229571] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1108.229832] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd7a513e-ea7a-4618-83e8-4ed4147e8e8f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.238525] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1108.238688] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1108.239501] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1f7ad9c-3cf4-4feb-91df-2c23058f1dbc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.245477] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1108.245477] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527abac7-5257-6ff9-376c-8c0b4a8c550c" [ 1108.245477] env[61985]: _type = "Task" [ 1108.245477] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.253480] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527abac7-5257-6ff9-376c-8c0b4a8c550c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.330525] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.879s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.332356] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.553s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.333511] env[61985]: DEBUG nova.objects.instance [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lazy-loading 'resources' on Instance uuid 87006fbf-ea90-4d9a-88af-001de424ac14 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.356923] env[61985]: INFO nova.scheduler.client.report [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Deleted allocations for instance 926a2b59-8614-4e71-9d94-fd9feefdef46 [ 1108.392108] env[61985]: DEBUG oslo_concurrency.lockutils [req-5d308d8d-48e7-4909-82d7-179726d153df req-3ed2f6f4-740c-4035-bd38-cf7d7cb063e0 service nova] Releasing lock "refresh_cache-f45760fa-2298-422e-84ff-83f99b19627d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.477795] env[61985]: DEBUG nova.compute.manager [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1108.480286] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936540, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.481196] env[61985]: DEBUG nova.network.neutron [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Updated VIF entry in instance network info cache for port 188cbfa9-832a-4537-ba42-d51325778b56. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1108.481589] env[61985]: DEBUG nova.network.neutron [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Updating instance_info_cache with network_info: [{"id": "0aa5789a-a605-4e56-aa4e-0402756b8569", "address": "fa:16:3e:fa:b3:fa", "network": {"id": "1f73c5a9-e080-48bf-922d-8d4739af644c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-367732363", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.93", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0aa5789a-a6", "ovs_interfaceid": "0aa5789a-a605-4e56-aa4e-0402756b8569", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "188cbfa9-832a-4537-ba42-d51325778b56", "address": "fa:16:3e:48:62:16", "network": {"id": "21e4b9d5-4291-4813-bbc7-020d922cff31", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1313355272", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.80", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap188cbfa9-83", "ovs_interfaceid": "188cbfa9-832a-4537-ba42-d51325778b56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.512237] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1108.512514] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1108.512681] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1108.512876] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1108.513044] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1108.513209] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1108.513426] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1108.513593] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1108.513781] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1108.513983] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1108.514189] env[61985]: DEBUG nova.virt.hardware [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1108.515517] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb847388-5527-4790-8409-7016b2156457 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.523820] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2dd4e7-9ca9-41c1-9550-e1a324fbfb73 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.756590] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527abac7-5257-6ff9-376c-8c0b4a8c550c, 'name': SearchDatastore_Task, 'duration_secs': 0.018863} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.757418] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4deed889-d463-47e8-a16f-c05d2d512aa9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.763310] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1108.763310] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b82576-c003-7e36-a982-58218b190cdd" [ 1108.763310] env[61985]: _type = "Task" [ 1108.763310] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.772016] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b82576-c003-7e36-a982-58218b190cdd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.829802] env[61985]: DEBUG nova.compute.manager [req-89fcd3c3-1654-497d-b102-4dd279b5ce60 req-b3a9d7fb-f72d-44bb-adcc-a37e17e9cb11 service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Received event network-vif-plugged-ecf04033-48b6-4b81-ab17-57a89dc0be60 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1108.830199] env[61985]: DEBUG oslo_concurrency.lockutils [req-89fcd3c3-1654-497d-b102-4dd279b5ce60 req-b3a9d7fb-f72d-44bb-adcc-a37e17e9cb11 service nova] Acquiring lock "ef330828-c218-4b80-8b1a-6813bdb67c4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.830708] env[61985]: DEBUG oslo_concurrency.lockutils [req-89fcd3c3-1654-497d-b102-4dd279b5ce60 req-b3a9d7fb-f72d-44bb-adcc-a37e17e9cb11 service nova] Lock "ef330828-c218-4b80-8b1a-6813bdb67c4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.830708] env[61985]: DEBUG oslo_concurrency.lockutils [req-89fcd3c3-1654-497d-b102-4dd279b5ce60 req-b3a9d7fb-f72d-44bb-adcc-a37e17e9cb11 service nova] Lock "ef330828-c218-4b80-8b1a-6813bdb67c4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.830933] env[61985]: DEBUG nova.compute.manager [req-89fcd3c3-1654-497d-b102-4dd279b5ce60 req-b3a9d7fb-f72d-44bb-adcc-a37e17e9cb11 service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] No waiting events found dispatching network-vif-plugged-ecf04033-48b6-4b81-ab17-57a89dc0be60 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1108.831119] env[61985]: WARNING nova.compute.manager [req-89fcd3c3-1654-497d-b102-4dd279b5ce60 req-b3a9d7fb-f72d-44bb-adcc-a37e17e9cb11 service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Received unexpected event network-vif-plugged-ecf04033-48b6-4b81-ab17-57a89dc0be60 for instance with vm_state building and task_state spawning. [ 1108.868203] env[61985]: DEBUG oslo_concurrency.lockutils [None req-eb11fcfe-565a-4242-8324-6750a0bf97f2 tempest-ServerDiskConfigTestJSON-1445928076 tempest-ServerDiskConfigTestJSON-1445928076-project-member] Lock "926a2b59-8614-4e71-9d94-fd9feefdef46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.716s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.943729] env[61985]: DEBUG nova.network.neutron [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Successfully updated port: ecf04033-48b6-4b81-ab17-57a89dc0be60 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1108.980425] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936540, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517973} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.980425] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 93dd0394-d7e1-42d5-a58f-fcd621c044b3/93dd0394-d7e1-42d5-a58f-fcd621c044b3.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1108.980531] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1108.983805] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f782d88-444b-40c0-bb3d-8eaee8aaf916 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.984439] env[61985]: DEBUG oslo_concurrency.lockutils [req-b5badf96-344a-4ac8-8f28-46b255e81c46 req-be364057-5c18-4e4e-b819-32e48c767505 service nova] Releasing lock "refresh_cache-93dd0394-d7e1-42d5-a58f-fcd621c044b3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.988026] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1108.988026] env[61985]: value = "task-936544" [ 1108.988026] env[61985]: _type = "Task" [ 1108.988026] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.000045] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936544, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.055784] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a7af6f3-72f0-48f3-a13b-52f977db4124 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.060114] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9923c06-badb-4476-963d-50618fb64710 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.082418] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8c97a3-735c-41bf-8793-2ff0289d70ed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.086529] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a94e79-81f3-4598-a780-f18ee69666f7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.094033] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance '885a36d9-24c6-407a-9d6b-01cd97e90674' progress to 83 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1109.123691] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0d3589-3f66-44ad-9180-4a37bb0e654f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.133870] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-615ca248-0967-4dee-a2e3-52e3ff9df9e5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.146051] env[61985]: DEBUG nova.compute.provider_tree [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.273825] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b82576-c003-7e36-a982-58218b190cdd, 'name': SearchDatastore_Task, 'duration_secs': 0.01} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.274107] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.274381] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] f45760fa-2298-422e-84ff-83f99b19627d/f45760fa-2298-422e-84ff-83f99b19627d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1109.274634] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56e06f1e-c71e-41f7-973f-72404b3e4b11 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.280457] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1109.280457] env[61985]: value = "task-936545" [ 1109.280457] env[61985]: _type = "Task" [ 1109.280457] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.288043] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.442577] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquiring lock "refresh_cache-ef330828-c218-4b80-8b1a-6813bdb67c4d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.442728] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquired lock "refresh_cache-ef330828-c218-4b80-8b1a-6813bdb67c4d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.442921] env[61985]: DEBUG nova.network.neutron [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1109.499919] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936544, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072091} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.500311] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1109.501136] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f714e421-26a9-47c6-9cbc-b0b779d021d9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.536741] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 93dd0394-d7e1-42d5-a58f-fcd621c044b3/93dd0394-d7e1-42d5-a58f-fcd621c044b3.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1109.537689] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6489088-1056-48da-9799-004bb2d33119 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.558289] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1109.558289] env[61985]: value = "task-936546" [ 1109.558289] env[61985]: _type = "Task" [ 1109.558289] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.566725] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936546, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.626528] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1109.627096] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-afc84f69-5f30-4d32-8351-b7564e110613 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.636313] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1109.636313] env[61985]: value = "task-936547" [ 1109.636313] env[61985]: _type = "Task" [ 1109.636313] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.650703] env[61985]: DEBUG nova.scheduler.client.report [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1109.656906] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936547, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.791608] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936545, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.995373] env[61985]: DEBUG nova.network.neutron [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1110.072242] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936546, 'name': ReconfigVM_Task, 'duration_secs': 0.423898} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.072611] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 93dd0394-d7e1-42d5-a58f-fcd621c044b3/93dd0394-d7e1-42d5-a58f-fcd621c044b3.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.073577] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e17c050b-90e6-4ac2-ba08-89431e18ef77 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.080893] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1110.080893] env[61985]: value = "task-936548" [ 1110.080893] env[61985]: _type = "Task" [ 1110.080893] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.095264] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936548, 'name': Rename_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.145683] env[61985]: DEBUG oslo_vmware.api [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936547, 'name': PowerOnVM_Task, 'duration_secs': 0.484718} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.146042] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1110.146267] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-75352ebb-53d1-40af-b766-213bb84c6c3b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance '885a36d9-24c6-407a-9d6b-01cd97e90674' progress to 100 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1110.159080] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.164279] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.017s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.165142] env[61985]: INFO nova.compute.claims [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1110.194257] env[61985]: INFO nova.scheduler.client.report [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Deleted allocations for instance 87006fbf-ea90-4d9a-88af-001de424ac14 [ 1110.221540] env[61985]: DEBUG nova.network.neutron [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Updating instance_info_cache with network_info: [{"id": "ecf04033-48b6-4b81-ab17-57a89dc0be60", "address": "fa:16:3e:b0:9f:2a", "network": {"id": "099d4473-cf19-4ac5-812d-ac0d14421e45", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-219826301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fa7a4522375426f9f3cc76e7f9270a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapecf04033-48", "ovs_interfaceid": "ecf04033-48b6-4b81-ab17-57a89dc0be60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.296785] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936545, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551938} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.296785] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] f45760fa-2298-422e-84ff-83f99b19627d/f45760fa-2298-422e-84ff-83f99b19627d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1110.296785] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1110.296785] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9dda12b-0e51-4259-ba4d-0a26d4439dbd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.303601] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1110.303601] env[61985]: value = "task-936549" [ 1110.303601] env[61985]: _type = "Task" [ 1110.303601] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.311030] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936549, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.590718] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936548, 'name': Rename_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.703481] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5600dff4-d5f6-4ce4-9c11-22af06c43780 tempest-ServersNegativeTestJSON-534843328 tempest-ServersNegativeTestJSON-534843328-project-member] Lock "87006fbf-ea90-4d9a-88af-001de424ac14" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.089s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.724884] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Releasing lock "refresh_cache-ef330828-c218-4b80-8b1a-6813bdb67c4d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.725388] env[61985]: DEBUG nova.compute.manager [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Instance network_info: |[{"id": "ecf04033-48b6-4b81-ab17-57a89dc0be60", "address": "fa:16:3e:b0:9f:2a", "network": {"id": "099d4473-cf19-4ac5-812d-ac0d14421e45", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-219826301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fa7a4522375426f9f3cc76e7f9270a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapecf04033-48", "ovs_interfaceid": "ecf04033-48b6-4b81-ab17-57a89dc0be60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1110.726204] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:9f:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bab6a6c3-1c5c-4776-b21b-dec21196d702', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ecf04033-48b6-4b81-ab17-57a89dc0be60', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1110.734103] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Creating folder: Project (8fa7a4522375426f9f3cc76e7f9270a5). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1110.734704] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-82b3dbe3-94ef-46f8-beee-c004e93306cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.746863] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Created folder: Project (8fa7a4522375426f9f3cc76e7f9270a5) in parent group-v211285. [ 1110.747473] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Creating folder: Instances. Parent ref: group-v211479. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1110.748142] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-147e85e8-803e-47bc-b436-17ff53d93496 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.761020] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Created folder: Instances in parent group-v211479. [ 1110.761020] env[61985]: DEBUG oslo.service.loopingcall [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.761020] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1110.761020] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ada9e957-e6ba-445d-9da4-d6bc40e7fe2d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.781317] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1110.781317] env[61985]: value = "task-936553" [ 1110.781317] env[61985]: _type = "Task" [ 1110.781317] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.792500] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936553, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.813394] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936549, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064688} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.813845] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1110.814740] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de1b2ec-aa0a-4698-a158-5787def0a7d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.838207] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] f45760fa-2298-422e-84ff-83f99b19627d/f45760fa-2298-422e-84ff-83f99b19627d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1110.838809] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-235eb923-2297-45d0-a753-b904b26ca4ca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.859833] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1110.859833] env[61985]: value = "task-936554" [ 1110.859833] env[61985]: _type = "Task" [ 1110.859833] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.868069] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936554, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.870263] env[61985]: DEBUG nova.compute.manager [req-92ba0f20-3d83-469c-8835-539fe8b6300e req-37d3a54f-e8dc-4be6-b519-d551bc4e13e7 service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Received event network-changed-ecf04033-48b6-4b81-ab17-57a89dc0be60 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1110.870507] env[61985]: DEBUG nova.compute.manager [req-92ba0f20-3d83-469c-8835-539fe8b6300e req-37d3a54f-e8dc-4be6-b519-d551bc4e13e7 service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Refreshing instance network info cache due to event network-changed-ecf04033-48b6-4b81-ab17-57a89dc0be60. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1110.870698] env[61985]: DEBUG oslo_concurrency.lockutils [req-92ba0f20-3d83-469c-8835-539fe8b6300e req-37d3a54f-e8dc-4be6-b519-d551bc4e13e7 service nova] Acquiring lock "refresh_cache-ef330828-c218-4b80-8b1a-6813bdb67c4d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.870856] env[61985]: DEBUG oslo_concurrency.lockutils [req-92ba0f20-3d83-469c-8835-539fe8b6300e req-37d3a54f-e8dc-4be6-b519-d551bc4e13e7 service nova] Acquired lock "refresh_cache-ef330828-c218-4b80-8b1a-6813bdb67c4d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.871569] env[61985]: DEBUG nova.network.neutron [req-92ba0f20-3d83-469c-8835-539fe8b6300e req-37d3a54f-e8dc-4be6-b519-d551bc4e13e7 service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Refreshing network info cache for port ecf04033-48b6-4b81-ab17-57a89dc0be60 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1111.092326] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936548, 'name': Rename_Task, 'duration_secs': 0.977651} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.092700] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1111.093929] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8097e4cd-fefb-4939-97a3-3f8a6f579ef1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.099737] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1111.099737] env[61985]: value = "task-936555" [ 1111.099737] env[61985]: _type = "Task" [ 1111.099737] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.107685] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936555, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.294073] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936553, 'name': CreateVM_Task, 'duration_secs': 0.39246} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.294388] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1111.295979] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.296217] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.296548] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1111.296815] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d7615c4-ee3f-4f86-88df-aa94c32efb47 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.304129] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for the task: (returnval){ [ 1111.304129] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5291f355-3b1d-00e4-72ab-f09da215b336" [ 1111.304129] env[61985]: _type = "Task" [ 1111.304129] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.311962] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5291f355-3b1d-00e4-72ab-f09da215b336, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.369648] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936554, 'name': ReconfigVM_Task, 'duration_secs': 0.383228} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.370174] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Reconfigured VM instance instance-00000068 to attach disk [datastore2] f45760fa-2298-422e-84ff-83f99b19627d/f45760fa-2298-422e-84ff-83f99b19627d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1111.370807] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc2e2a1d-780f-4054-8198-bfe603bf2142 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.378260] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1111.378260] env[61985]: value = "task-936556" [ 1111.378260] env[61985]: _type = "Task" [ 1111.378260] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.378958] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35471328-75e4-4a75-9db7-2140f3f16007 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.389737] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936556, 'name': Rename_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.393750] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3816b9-bcf6-43d9-9759-0d34855fbb38 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.197095] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68bbaa43-ad91-4c53-99da-4bcfe813bd2f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.215473] env[61985]: DEBUG oslo_vmware.api [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936555, 'name': PowerOnVM_Task, 'duration_secs': 0.479314} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.215738] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936556, 'name': Rename_Task, 'duration_secs': 0.169328} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.215951] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5291f355-3b1d-00e4-72ab-f09da215b336, 'name': SearchDatastore_Task, 'duration_secs': 0.021958} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.217782] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1112.218089] env[61985]: INFO nova.compute.manager [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Took 12.12 seconds to spawn the instance on the hypervisor. [ 1112.218237] env[61985]: DEBUG nova.compute.manager [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1112.218562] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1112.218775] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.218991] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1112.219240] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.219390] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.219568] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1112.221283] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bba53b-36a2-4e06-88cb-b44b8858493f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.228299] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade4637e-310d-4234-a540-a7baa18ff6ce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.231114] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0f796a6-61cb-4882-8e5b-3f742add2dc3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.232449] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c8495b2-c523-45b0-84d1-aa2c298d2767 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.245846] env[61985]: DEBUG nova.compute.provider_tree [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.250129] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1112.250271] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1112.253577] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1112.253577] env[61985]: value = "task-936557" [ 1112.253577] env[61985]: _type = "Task" [ 1112.253577] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.253955] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7aa9d583-8b8b-445a-a317-1db89e378b59 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.266057] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for the task: (returnval){ [ 1112.266057] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523e51ed-2a57-7906-52b1-395961af0244" [ 1112.266057] env[61985]: _type = "Task" [ 1112.266057] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.269531] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936557, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.278501] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523e51ed-2a57-7906-52b1-395961af0244, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.405062] env[61985]: DEBUG nova.network.neutron [req-92ba0f20-3d83-469c-8835-539fe8b6300e req-37d3a54f-e8dc-4be6-b519-d551bc4e13e7 service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Updated VIF entry in instance network info cache for port ecf04033-48b6-4b81-ab17-57a89dc0be60. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1112.405390] env[61985]: DEBUG nova.network.neutron [req-92ba0f20-3d83-469c-8835-539fe8b6300e req-37d3a54f-e8dc-4be6-b519-d551bc4e13e7 service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Updating instance_info_cache with network_info: [{"id": "ecf04033-48b6-4b81-ab17-57a89dc0be60", "address": "fa:16:3e:b0:9f:2a", "network": {"id": "099d4473-cf19-4ac5-812d-ac0d14421e45", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-219826301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fa7a4522375426f9f3cc76e7f9270a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapecf04033-48", "ovs_interfaceid": "ecf04033-48b6-4b81-ab17-57a89dc0be60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.478440] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "f2bd7642-e20c-408b-9a7f-73428b89328a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.478992] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "f2bd7642-e20c-408b-9a7f-73428b89328a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.679819] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1112.680138] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211478', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'name': 'volume-9d006806-6309-415f-b72a-5e327ec77248', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '036fae78-2af0-49f1-85fa-245c29f35ebc', 'attached_at': '', 'detached_at': '', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'serial': '9d006806-6309-415f-b72a-5e327ec77248'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1112.681216] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab22ab7-c0d4-40ad-9404-82fec1b29616 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.699993] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc9f697-dee1-474a-8799-99ea71ddbc61 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.725666] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-9d006806-6309-415f-b72a-5e327ec77248/volume-9d006806-6309-415f-b72a-5e327ec77248.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.726027] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7590d965-f825-463c-a569-950bd0805eec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.747757] env[61985]: DEBUG oslo_vmware.api [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1112.747757] env[61985]: value = "task-936558" [ 1112.747757] env[61985]: _type = "Task" [ 1112.747757] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.755626] env[61985]: DEBUG nova.scheduler.client.report [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.776303] env[61985]: DEBUG oslo_vmware.api [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.776853] env[61985]: INFO nova.compute.manager [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Took 23.21 seconds to build instance. [ 1112.793230] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523e51ed-2a57-7906-52b1-395961af0244, 'name': SearchDatastore_Task, 'duration_secs': 0.017957} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.793230] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936557, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.794235] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45deb0fe-51e9-4f76-9e79-a90fc482c518 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.801014] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for the task: (returnval){ [ 1112.801014] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ce8a95-2af0-d447-42d8-61efeb575c15" [ 1112.801014] env[61985]: _type = "Task" [ 1112.801014] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.810467] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ce8a95-2af0-d447-42d8-61efeb575c15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.908156] env[61985]: DEBUG oslo_concurrency.lockutils [req-92ba0f20-3d83-469c-8835-539fe8b6300e req-37d3a54f-e8dc-4be6-b519-d551bc4e13e7 service nova] Releasing lock "refresh_cache-ef330828-c218-4b80-8b1a-6813bdb67c4d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.937889] env[61985]: DEBUG nova.network.neutron [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Port a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1112.938249] env[61985]: DEBUG oslo_concurrency.lockutils [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.938479] env[61985]: DEBUG oslo_concurrency.lockutils [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.938730] env[61985]: DEBUG nova.network.neutron [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1112.983873] env[61985]: DEBUG nova.compute.manager [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1113.261133] env[61985]: DEBUG oslo_vmware.api [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.267361] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.104s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.267948] env[61985]: DEBUG nova.compute.manager [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1113.273642] env[61985]: DEBUG oslo_vmware.api [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936557, 'name': PowerOnVM_Task, 'duration_secs': 0.675544} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.274118] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1113.274333] env[61985]: INFO nova.compute.manager [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Took 8.94 seconds to spawn the instance on the hypervisor. [ 1113.274516] env[61985]: DEBUG nova.compute.manager [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1113.275318] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6297d439-2089-4635-b302-e6da62d7dc3f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.284362] env[61985]: DEBUG oslo_concurrency.lockutils [None req-915a08d5-1a73-43e0-a8de-56dbcb83cbd2 tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.727s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.313677] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ce8a95-2af0-d447-42d8-61efeb575c15, 'name': SearchDatastore_Task, 'duration_secs': 0.010802} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.313929] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.314205] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] ef330828-c218-4b80-8b1a-6813bdb67c4d/ef330828-c218-4b80-8b1a-6813bdb67c4d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1113.314461] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4810500-e199-44e9-a17e-82cf72bcc52f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.322448] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for the task: (returnval){ [ 1113.322448] env[61985]: value = "task-936559" [ 1113.322448] env[61985]: _type = "Task" [ 1113.322448] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.330475] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936559, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.428809] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.429496] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.429496] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.429671] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.429822] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.432647] env[61985]: INFO nova.compute.manager [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Terminating instance [ 1113.434591] env[61985]: DEBUG nova.compute.manager [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1113.434792] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1113.435881] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602df391-1888-4b99-ab24-4bce5aff3f60 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.445772] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1113.446402] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab446d83-418b-4a68-a8ab-1c6c75144e28 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.454674] env[61985]: DEBUG oslo_vmware.api [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1113.454674] env[61985]: value = "task-936560" [ 1113.454674] env[61985]: _type = "Task" [ 1113.454674] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.464518] env[61985]: DEBUG oslo_vmware.api [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.510330] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.510573] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.512148] env[61985]: INFO nova.compute.claims [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1113.692845] env[61985]: DEBUG nova.network.neutron [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance_info_cache with network_info: [{"id": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "address": "fa:16:3e:e1:d0:1f", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64c525f-b3", "ovs_interfaceid": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.760671] env[61985]: DEBUG oslo_vmware.api [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936558, 'name': ReconfigVM_Task, 'duration_secs': 0.663708} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.761009] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-9d006806-6309-415f-b72a-5e327ec77248/volume-9d006806-6309-415f-b72a-5e327ec77248.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1113.766113] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6dd2ad72-b4ed-432e-983f-7e7e14bbff0e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.777801] env[61985]: DEBUG nova.compute.utils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1113.779329] env[61985]: DEBUG nova.compute.manager [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1113.779507] env[61985]: DEBUG nova.network.neutron [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1113.793856] env[61985]: DEBUG oslo_vmware.api [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1113.793856] env[61985]: value = "task-936561" [ 1113.793856] env[61985]: _type = "Task" [ 1113.793856] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.794179] env[61985]: INFO nova.compute.manager [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Took 14.69 seconds to build instance. [ 1113.806685] env[61985]: DEBUG oslo_vmware.api [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936561, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.828750] env[61985]: DEBUG nova.policy [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'faa19f385cd0426d8565e31e286800d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5625013599b44418bd56eb604e14be58', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1113.837170] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936559, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.966712] env[61985]: DEBUG oslo_vmware.api [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936560, 'name': PowerOffVM_Task, 'duration_secs': 0.273009} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.966902] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1113.967117] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1113.967400] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-114e79db-5659-48b7-9d7a-700a3c95d884 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.123363] env[61985]: DEBUG nova.network.neutron [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Successfully created port: 7303cfe7-ecb8-4ee1-9b69-414635f73c1b {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1114.128198] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1114.128534] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1114.128818] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Deleting the datastore file [datastore1] 93dd0394-d7e1-42d5-a58f-fcd621c044b3 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1114.129206] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3c220a3-2aca-4577-a36f-fac8e5ef387b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.138169] env[61985]: DEBUG oslo_vmware.api [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for the task: (returnval){ [ 1114.138169] env[61985]: value = "task-936563" [ 1114.138169] env[61985]: _type = "Task" [ 1114.138169] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.149486] env[61985]: DEBUG oslo_vmware.api [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936563, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.195922] env[61985]: DEBUG oslo_concurrency.lockutils [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.284063] env[61985]: DEBUG nova.compute.manager [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1114.296930] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d2a3c228-734f-44c4-80c2-25076f31abd1 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "f45760fa-2298-422e-84ff-83f99b19627d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.202s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.306530] env[61985]: DEBUG oslo_vmware.api [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936561, 'name': ReconfigVM_Task, 'duration_secs': 0.185919} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.306828] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211478', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'name': 'volume-9d006806-6309-415f-b72a-5e327ec77248', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '036fae78-2af0-49f1-85fa-245c29f35ebc', 'attached_at': '', 'detached_at': '', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'serial': '9d006806-6309-415f-b72a-5e327ec77248'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1114.334956] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936559, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.74651} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.335251] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] ef330828-c218-4b80-8b1a-6813bdb67c4d/ef330828-c218-4b80-8b1a-6813bdb67c4d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1114.335567] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1114.335992] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1ed1c4a-040d-4dda-964c-4ee37194ccc2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.344408] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for the task: (returnval){ [ 1114.344408] env[61985]: value = "task-936564" [ 1114.344408] env[61985]: _type = "Task" [ 1114.344408] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.352780] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936564, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.651266] env[61985]: DEBUG oslo_vmware.api [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936563, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.699937] env[61985]: DEBUG nova.compute.manager [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61985) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1114.726791] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bbf2cb5-0788-4ef8-b31f-86ce8559b7fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.737011] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05eaa3a7-72ca-4d9c-aa7b-793baf3a48ea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.768392] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeedc33c-cc31-47f3-a171-3ccbf66f92b1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.778327] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea664394-a6b3-4d9f-89fa-4a12dda9752b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.798085] env[61985]: DEBUG nova.compute.provider_tree [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.853879] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936564, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073723} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.854188] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1114.854966] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151769f3-306f-47e2-b8e4-f5e42dab19ec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.882404] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] ef330828-c218-4b80-8b1a-6813bdb67c4d/ef330828-c218-4b80-8b1a-6813bdb67c4d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1114.882910] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd9bf3f5-c316-4cf3-a8f7-d1d68db7758f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.907947] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for the task: (returnval){ [ 1114.907947] env[61985]: value = "task-936565" [ 1114.907947] env[61985]: _type = "Task" [ 1114.907947] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.918804] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936565, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.988059] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "f45760fa-2298-422e-84ff-83f99b19627d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.988481] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "f45760fa-2298-422e-84ff-83f99b19627d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.988741] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "f45760fa-2298-422e-84ff-83f99b19627d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.988999] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "f45760fa-2298-422e-84ff-83f99b19627d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.989199] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "f45760fa-2298-422e-84ff-83f99b19627d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.991750] env[61985]: INFO nova.compute.manager [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Terminating instance [ 1114.994119] env[61985]: DEBUG nova.compute.manager [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1114.994341] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1114.995188] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7be75f-96ee-49be-a8e6-161c556d48bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.003996] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1115.004274] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-621d748e-50de-467e-92bd-069320aa860d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.014426] env[61985]: DEBUG oslo_vmware.api [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1115.014426] env[61985]: value = "task-936566" [ 1115.014426] env[61985]: _type = "Task" [ 1115.014426] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.023411] env[61985]: DEBUG oslo_vmware.api [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936566, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.151039] env[61985]: DEBUG oslo_vmware.api [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Task: {'id': task-936563, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.794697} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.151039] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1115.151224] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1115.151350] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1115.151492] env[61985]: INFO nova.compute.manager [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Took 1.72 seconds to destroy the instance on the hypervisor. [ 1115.151745] env[61985]: DEBUG oslo.service.loopingcall [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.151943] env[61985]: DEBUG nova.compute.manager [-] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1115.152060] env[61985]: DEBUG nova.network.neutron [-] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1115.302111] env[61985]: DEBUG nova.compute.manager [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1115.304733] env[61985]: DEBUG nova.scheduler.client.report [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1115.337143] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1115.337389] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1115.337565] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1115.337848] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1115.338146] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1115.338415] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1115.338745] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1115.338991] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1115.339222] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1115.339403] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1115.339585] env[61985]: DEBUG nova.virt.hardware [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1115.341039] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d01c9a5-3897-4702-a93e-c9e845270643 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.349948] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1be5f0-d773-4d98-a529-823e36308564 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.403543] env[61985]: DEBUG nova.objects.instance [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lazy-loading 'flavor' on Instance uuid 036fae78-2af0-49f1-85fa-245c29f35ebc {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.417537] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936565, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.512534] env[61985]: DEBUG nova.compute.manager [req-ae2fb897-4dd7-4697-a0ba-701a1213697d req-17be15f9-081f-4703-93ff-8c3c6e92bcc3 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Received event network-vif-deleted-0aa5789a-a605-4e56-aa4e-0402756b8569 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1115.512745] env[61985]: INFO nova.compute.manager [req-ae2fb897-4dd7-4697-a0ba-701a1213697d req-17be15f9-081f-4703-93ff-8c3c6e92bcc3 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Neutron deleted interface 0aa5789a-a605-4e56-aa4e-0402756b8569; detaching it from the instance and deleting it from the info cache [ 1115.513018] env[61985]: DEBUG nova.network.neutron [req-ae2fb897-4dd7-4697-a0ba-701a1213697d req-17be15f9-081f-4703-93ff-8c3c6e92bcc3 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Updating instance_info_cache with network_info: [{"id": "188cbfa9-832a-4537-ba42-d51325778b56", "address": "fa:16:3e:48:62:16", "network": {"id": "21e4b9d5-4291-4813-bbc7-020d922cff31", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1313355272", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.80", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "60ecaef69401418eac6c655c1b7df5be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "504a3211-8288-4b87-a52a-590466d80a49", "external-id": "nsx-vlan-transportzone-469", "segmentation_id": 469, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap188cbfa9-83", "ovs_interfaceid": "188cbfa9-832a-4537-ba42-d51325778b56", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.525739] env[61985]: DEBUG oslo_vmware.api [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936566, 'name': PowerOffVM_Task, 'duration_secs': 0.215976} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.526018] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1115.530020] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1115.530020] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8e1783f-ef74-4ddf-8b0d-34bb4b9329ef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.626623] env[61985]: DEBUG nova.compute.manager [req-efa44178-0117-4bc0-9d28-02e296c45adb req-3a355f1e-fd30-4835-abfb-17a8c46a5503 service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Received event network-vif-plugged-7303cfe7-ecb8-4ee1-9b69-414635f73c1b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1115.626847] env[61985]: DEBUG oslo_concurrency.lockutils [req-efa44178-0117-4bc0-9d28-02e296c45adb req-3a355f1e-fd30-4835-abfb-17a8c46a5503 service nova] Acquiring lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.627133] env[61985]: DEBUG oslo_concurrency.lockutils [req-efa44178-0117-4bc0-9d28-02e296c45adb req-3a355f1e-fd30-4835-abfb-17a8c46a5503 service nova] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.627260] env[61985]: DEBUG oslo_concurrency.lockutils [req-efa44178-0117-4bc0-9d28-02e296c45adb req-3a355f1e-fd30-4835-abfb-17a8c46a5503 service nova] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.627430] env[61985]: DEBUG nova.compute.manager [req-efa44178-0117-4bc0-9d28-02e296c45adb req-3a355f1e-fd30-4835-abfb-17a8c46a5503 service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] No waiting events found dispatching network-vif-plugged-7303cfe7-ecb8-4ee1-9b69-414635f73c1b {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1115.627598] env[61985]: WARNING nova.compute.manager [req-efa44178-0117-4bc0-9d28-02e296c45adb req-3a355f1e-fd30-4835-abfb-17a8c46a5503 service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Received unexpected event network-vif-plugged-7303cfe7-ecb8-4ee1-9b69-414635f73c1b for instance with vm_state building and task_state spawning. [ 1115.814420] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.303s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.814420] env[61985]: DEBUG nova.compute.manager [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1115.823726] env[61985]: INFO nova.compute.manager [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Rebuilding instance [ 1115.831793] env[61985]: DEBUG oslo_concurrency.lockutils [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.832203] env[61985]: DEBUG oslo_concurrency.lockutils [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.882069] env[61985]: DEBUG nova.compute.manager [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1115.882975] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-febe4305-07b7-4d42-bb68-fb647e6ac2d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.911494] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1b6f8b5a-de0b-4dd8-8086-5a79c193a9eb tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.832s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.918131] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936565, 'name': ReconfigVM_Task, 'duration_secs': 0.907279} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.918307] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Reconfigured VM instance instance-00000069 to attach disk [datastore2] ef330828-c218-4b80-8b1a-6813bdb67c4d/ef330828-c218-4b80-8b1a-6813bdb67c4d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1115.918878] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf5d6274-63d2-4e95-8562-3c94a8b110df {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.928440] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for the task: (returnval){ [ 1115.928440] env[61985]: value = "task-936568" [ 1115.928440] env[61985]: _type = "Task" [ 1115.928440] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.941933] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936568, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.943288] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1115.943590] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1115.943663] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleting the datastore file [datastore2] f45760fa-2298-422e-84ff-83f99b19627d {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1115.943936] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b7a37f3-56eb-4245-abac-be079eeb8864 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.951491] env[61985]: DEBUG oslo_vmware.api [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1115.951491] env[61985]: value = "task-936569" [ 1115.951491] env[61985]: _type = "Task" [ 1115.951491] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.961322] env[61985]: DEBUG oslo_vmware.api [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936569, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.016235] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f18852d-a4de-4e6b-9bae-1018c2ce17d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.026792] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b71b1ce-6e7b-4ed9-b908-8ad45b0a388f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.042310] env[61985]: DEBUG oslo_concurrency.lockutils [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "9079e605-e865-4eb1-9857-03769e4b5ba6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.042470] env[61985]: DEBUG oslo_concurrency.lockutils [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.063240] env[61985]: DEBUG nova.compute.manager [req-ae2fb897-4dd7-4697-a0ba-701a1213697d req-17be15f9-081f-4703-93ff-8c3c6e92bcc3 service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Detach interface failed, port_id=0aa5789a-a605-4e56-aa4e-0402756b8569, reason: Instance 93dd0394-d7e1-42d5-a58f-fcd621c044b3 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1116.200591] env[61985]: DEBUG nova.network.neutron [-] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.214051] env[61985]: DEBUG nova.network.neutron [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Successfully updated port: 7303cfe7-ecb8-4ee1-9b69-414635f73c1b {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1116.319943] env[61985]: DEBUG nova.compute.utils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1116.324592] env[61985]: DEBUG nova.compute.manager [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Not allocating networking since 'none' was specified. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1116.334379] env[61985]: DEBUG nova.objects.instance [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'migration_context' on Instance uuid 885a36d9-24c6-407a-9d6b-01cd97e90674 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.395147] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1116.395469] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-821d8edd-5fb9-45d8-9b24-e12a53ed8636 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.404667] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1116.404667] env[61985]: value = "task-936570" [ 1116.404667] env[61985]: _type = "Task" [ 1116.404667] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.416270] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936570, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.439029] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936568, 'name': Rename_Task, 'duration_secs': 0.16671} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.439480] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1116.439736] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7183733-392e-44b9-8fc6-05c3d23eb9b1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.447015] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for the task: (returnval){ [ 1116.447015] env[61985]: value = "task-936571" [ 1116.447015] env[61985]: _type = "Task" [ 1116.447015] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.458121] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936571, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.463675] env[61985]: DEBUG oslo_vmware.api [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936569, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125915} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.463930] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1116.464150] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1116.464325] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1116.464505] env[61985]: INFO nova.compute.manager [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Took 1.47 seconds to destroy the instance on the hypervisor. [ 1116.464751] env[61985]: DEBUG oslo.service.loopingcall [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1116.464958] env[61985]: DEBUG nova.compute.manager [-] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1116.465079] env[61985]: DEBUG nova.network.neutron [-] [instance: f45760fa-2298-422e-84ff-83f99b19627d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1116.545432] env[61985]: DEBUG nova.compute.utils [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1116.705628] env[61985]: INFO nova.compute.manager [-] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Took 1.55 seconds to deallocate network for instance. [ 1116.716184] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.717168] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.717400] env[61985]: DEBUG nova.network.neutron [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1116.823271] env[61985]: DEBUG nova.compute.manager [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1116.914186] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936570, 'name': PowerOffVM_Task, 'duration_secs': 0.346068} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.916581] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1116.962193] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936571, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.974736] env[61985]: INFO nova.compute.manager [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Detaching volume 9d006806-6309-415f-b72a-5e327ec77248 [ 1117.008818] env[61985]: INFO nova.virt.block_device [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Attempting to driver detach volume 9d006806-6309-415f-b72a-5e327ec77248 from mountpoint /dev/sdb [ 1117.009094] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1117.009300] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211478', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'name': 'volume-9d006806-6309-415f-b72a-5e327ec77248', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '036fae78-2af0-49f1-85fa-245c29f35ebc', 'attached_at': '', 'detached_at': '', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'serial': '9d006806-6309-415f-b72a-5e327ec77248'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1117.010237] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2f68be-ae43-45d3-a1a3-0702ad5c5fdc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.014093] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469985e2-fa57-4f62-85f0-d754c7583951 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.037744] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338061e4-fb6e-4a89-abd2-06caa2944e50 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.041032] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4ed19f-7a12-4be6-a413-ea394f129728 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.048840] env[61985]: DEBUG oslo_concurrency.lockutils [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.075614] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae092853-47ef-43f7-ac0a-2a43637c9646 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.078633] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45a3b16-b8a6-4b51-b7a2-a558165793e0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.103535] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244bed78-2eb9-41e1-812d-55569a6b8694 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.108186] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba394acf-d0b9-45a5-b5a3-b388d5b4b355 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.122694] env[61985]: DEBUG nova.compute.provider_tree [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.134514] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] The volume has not been displaced from its original location: [datastore2] volume-9d006806-6309-415f-b72a-5e327ec77248/volume-9d006806-6309-415f-b72a-5e327ec77248.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1117.139837] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1117.140973] env[61985]: DEBUG nova.scheduler.client.report [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.144203] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5b5ddb6-bcd4-4eee-8390-56cb46c4f37d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.164015] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1117.164015] env[61985]: value = "task-936572" [ 1117.164015] env[61985]: _type = "Task" [ 1117.164015] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.174359] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936572, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.213527] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.228914] env[61985]: DEBUG nova.network.neutron [-] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.248916] env[61985]: DEBUG nova.network.neutron [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1117.399269] env[61985]: DEBUG nova.network.neutron [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance_info_cache with network_info: [{"id": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "address": "fa:16:3e:d8:5d:8b", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7303cfe7-ec", "ovs_interfaceid": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.459444] env[61985]: DEBUG oslo_vmware.api [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936571, 'name': PowerOnVM_Task, 'duration_secs': 0.975546} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.459733] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1117.459980] env[61985]: INFO nova.compute.manager [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Took 8.98 seconds to spawn the instance on the hypervisor. [ 1117.460216] env[61985]: DEBUG nova.compute.manager [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1117.461021] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b61373-f3a0-41ed-8265-5624f46b352d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.619227] env[61985]: DEBUG nova.compute.manager [req-5b67e383-43b4-4abc-8180-b13bbfb1b377 req-2c4aaff8-145a-445b-8302-75815d24d97d service nova] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Received event network-vif-deleted-188cbfa9-832a-4537-ba42-d51325778b56 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1117.619293] env[61985]: DEBUG nova.compute.manager [req-5b67e383-43b4-4abc-8180-b13bbfb1b377 req-2c4aaff8-145a-445b-8302-75815d24d97d service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Received event network-changed-7303cfe7-ecb8-4ee1-9b69-414635f73c1b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1117.619441] env[61985]: DEBUG nova.compute.manager [req-5b67e383-43b4-4abc-8180-b13bbfb1b377 req-2c4aaff8-145a-445b-8302-75815d24d97d service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Refreshing instance network info cache due to event network-changed-7303cfe7-ecb8-4ee1-9b69-414635f73c1b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1117.619638] env[61985]: DEBUG oslo_concurrency.lockutils [req-5b67e383-43b4-4abc-8180-b13bbfb1b377 req-2c4aaff8-145a-445b-8302-75815d24d97d service nova] Acquiring lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.675442] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936572, 'name': ReconfigVM_Task, 'duration_secs': 0.230274} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.675892] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1117.682300] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82ccf879-2ff4-4da6-a98b-498aa152d78e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.700674] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1117.700674] env[61985]: value = "task-936573" [ 1117.700674] env[61985]: _type = "Task" [ 1117.700674] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.710649] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936573, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.732550] env[61985]: INFO nova.compute.manager [-] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Took 1.27 seconds to deallocate network for instance. [ 1117.835732] env[61985]: DEBUG nova.compute.manager [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1117.871613] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1117.871911] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1117.872126] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1117.872322] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1117.872480] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1117.872638] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1117.872927] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1117.873163] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1117.873368] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1117.873573] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1117.873885] env[61985]: DEBUG nova.virt.hardware [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1117.874897] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb314284-45ee-4c14-98db-bc7cd0f35478 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.885192] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532e6c6a-4305-4ea2-9c79-9511e348160a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.901589] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.902010] env[61985]: DEBUG nova.compute.manager [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Instance network_info: |[{"id": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "address": "fa:16:3e:d8:5d:8b", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7303cfe7-ec", "ovs_interfaceid": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1117.902609] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1117.908888] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Creating folder: Project (bed567b49f564ae58daee936346ed084). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1117.909344] env[61985]: DEBUG oslo_concurrency.lockutils [req-5b67e383-43b4-4abc-8180-b13bbfb1b377 req-2c4aaff8-145a-445b-8302-75815d24d97d service nova] Acquired lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.909592] env[61985]: DEBUG nova.network.neutron [req-5b67e383-43b4-4abc-8180-b13bbfb1b377 req-2c4aaff8-145a-445b-8302-75815d24d97d service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Refreshing network info cache for port 7303cfe7-ecb8-4ee1-9b69-414635f73c1b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1117.911116] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:5d:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7908211b-df93-467b-87a8-3c3d29b03de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7303cfe7-ecb8-4ee1-9b69-414635f73c1b', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1117.919247] env[61985]: DEBUG oslo.service.loopingcall [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.919475] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c3a801f7-859c-4222-b895-713afb7e6386 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.922250] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1117.922538] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-656bd3bb-b378-4900-8284-652f054c46a9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.951460] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1117.951460] env[61985]: value = "task-936575" [ 1117.951460] env[61985]: _type = "Task" [ 1117.951460] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.956449] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Created folder: Project (bed567b49f564ae58daee936346ed084) in parent group-v211285. [ 1117.956661] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Creating folder: Instances. Parent ref: group-v211482. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1117.957316] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5fcb339-80a3-48ed-9a7f-147b6bc84889 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.964623] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936575, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.969635] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Created folder: Instances in parent group-v211482. [ 1117.969972] env[61985]: DEBUG oslo.service.loopingcall [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.970258] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1117.970486] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f8ca6c1e-f075-4963-9c96-ca54824ad903 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.995686] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1117.995686] env[61985]: value = "task-936577" [ 1117.995686] env[61985]: _type = "Task" [ 1117.995686] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.000745] env[61985]: INFO nova.compute.manager [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Took 14.75 seconds to build instance. [ 1118.007230] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936577, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.101894] env[61985]: DEBUG oslo_concurrency.lockutils [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "9079e605-e865-4eb1-9857-03769e4b5ba6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.102249] env[61985]: DEBUG oslo_concurrency.lockutils [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.102505] env[61985]: INFO nova.compute.manager [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Attaching volume 03676853-7534-4379-bea5-3bf20b6efccd to /dev/sdb [ 1118.142347] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fafa89-0608-4b01-84a0-578ee373dfb5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.151306] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32fe6f4-4a37-4a22-b5c6-3fb32cdbd1f3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.165116] env[61985]: DEBUG oslo_concurrency.lockutils [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.333s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.174932] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.961s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.175336] env[61985]: DEBUG nova.objects.instance [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lazy-loading 'resources' on Instance uuid 93dd0394-d7e1-42d5-a58f-fcd621c044b3 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.176448] env[61985]: DEBUG nova.virt.block_device [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Updating existing volume attachment record: 1165df7b-8dc6-464e-8ee9-1285ca75c57f {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1118.217145] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936573, 'name': ReconfigVM_Task, 'duration_secs': 0.19976} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.217960] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211478', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'name': 'volume-9d006806-6309-415f-b72a-5e327ec77248', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '036fae78-2af0-49f1-85fa-245c29f35ebc', 'attached_at': '', 'detached_at': '', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'serial': '9d006806-6309-415f-b72a-5e327ec77248'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1118.240265] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.303797] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquiring lock "ef330828-c218-4b80-8b1a-6813bdb67c4d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.464389] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936575, 'name': CreateVM_Task, 'duration_secs': 0.417289} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.464871] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1118.465381] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.465638] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.466281] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1118.466281] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73b24a88-6198-4a86-9aa9-4866e676da9d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.471446] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1118.471446] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fa4b9f-e5a8-75e7-3300-c9113e7af069" [ 1118.471446] env[61985]: _type = "Task" [ 1118.471446] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.480029] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fa4b9f-e5a8-75e7-3300-c9113e7af069, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.505308] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6354e53b-c1e7-41ef-8743-337c9d5f2f0e tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lock "ef330828-c218-4b80-8b1a-6813bdb67c4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.269s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.505558] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936577, 'name': CreateVM_Task, 'duration_secs': 0.33413} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.507923] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lock "ef330828-c218-4b80-8b1a-6813bdb67c4d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.204s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.508182] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquiring lock "ef330828-c218-4b80-8b1a-6813bdb67c4d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.508435] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lock "ef330828-c218-4b80-8b1a-6813bdb67c4d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.508616] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lock "ef330828-c218-4b80-8b1a-6813bdb67c4d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.510184] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1118.510574] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.511300] env[61985]: INFO nova.compute.manager [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Terminating instance [ 1118.513102] env[61985]: DEBUG nova.compute.manager [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1118.513306] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1118.514076] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10b8fec-17c8-4394-a29f-e50163e4fa2a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.521744] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1118.521998] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff938203-9969-4089-b6ea-5b746a9c2839 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.529315] env[61985]: DEBUG oslo_vmware.api [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for the task: (returnval){ [ 1118.529315] env[61985]: value = "task-936579" [ 1118.529315] env[61985]: _type = "Task" [ 1118.529315] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.537404] env[61985]: DEBUG oslo_vmware.api [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936579, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.643876] env[61985]: DEBUG nova.network.neutron [req-5b67e383-43b4-4abc-8180-b13bbfb1b377 req-2c4aaff8-145a-445b-8302-75815d24d97d service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updated VIF entry in instance network info cache for port 7303cfe7-ecb8-4ee1-9b69-414635f73c1b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1118.644328] env[61985]: DEBUG nova.network.neutron [req-5b67e383-43b4-4abc-8180-b13bbfb1b377 req-2c4aaff8-145a-445b-8302-75815d24d97d service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance_info_cache with network_info: [{"id": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "address": "fa:16:3e:d8:5d:8b", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7303cfe7-ec", "ovs_interfaceid": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.865869] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a265a4-f4ac-4f6d-aef9-6f388b9c4a38 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.875087] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe8e216-eec7-4ddf-8c82-b58c6a5d511a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.905877] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48098fdc-f5bd-45af-a154-53c97873e899 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.913662] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58480b04-f7fc-4e8c-870b-5a830722af2e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.928400] env[61985]: DEBUG nova.compute.provider_tree [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.981241] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fa4b9f-e5a8-75e7-3300-c9113e7af069, 'name': SearchDatastore_Task, 'duration_secs': 0.010599} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.981533] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.981772] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1118.982021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.982176] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.982356] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1118.982878] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.982973] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1118.983169] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d75b2a69-2594-4f5d-948e-b4bdefe2d1c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.984828] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e6895fb-003e-4081-b626-e63ecf05ed6e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.990168] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1118.990168] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52db3acd-2dcc-91d1-92c8-1cc956e5817c" [ 1118.990168] env[61985]: _type = "Task" [ 1118.990168] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.993846] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1118.994073] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1118.994996] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c29b9d26-267f-4847-9151-9022aba46897 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.999816] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52db3acd-2dcc-91d1-92c8-1cc956e5817c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.003243] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1119.003243] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5219582b-89a0-44dc-1a63-40ba6744a442" [ 1119.003243] env[61985]: _type = "Task" [ 1119.003243] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.010892] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5219582b-89a0-44dc-1a63-40ba6744a442, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.039743] env[61985]: DEBUG oslo_vmware.api [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936579, 'name': PowerOffVM_Task, 'duration_secs': 0.172434} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.040046] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1119.040241] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1119.040495] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-592daa07-f9d9-4c8e-ae6c-aa56479ee422 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.114874] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1119.115146] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1119.115286] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Deleting the datastore file [datastore2] ef330828-c218-4b80-8b1a-6813bdb67c4d {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1119.115611] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30280937-86ca-4fdc-bb2b-69ca3cd28647 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.123664] env[61985]: DEBUG oslo_vmware.api [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for the task: (returnval){ [ 1119.123664] env[61985]: value = "task-936581" [ 1119.123664] env[61985]: _type = "Task" [ 1119.123664] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.132444] env[61985]: DEBUG oslo_vmware.api [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936581, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.147115] env[61985]: DEBUG oslo_concurrency.lockutils [req-5b67e383-43b4-4abc-8180-b13bbfb1b377 req-2c4aaff8-145a-445b-8302-75815d24d97d service nova] Releasing lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.147519] env[61985]: DEBUG nova.compute.manager [req-5b67e383-43b4-4abc-8180-b13bbfb1b377 req-2c4aaff8-145a-445b-8302-75815d24d97d service nova] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Received event network-vif-deleted-650759b1-a52a-4181-ac42-6744da8673c9 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1119.273463] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1119.274284] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21fef80d-2e6f-4fd4-9d0b-6b91dd418400 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.283728] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1119.283728] env[61985]: value = "task-936582" [ 1119.283728] env[61985]: _type = "Task" [ 1119.283728] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.294884] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1119.295051] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1119.295226] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211478', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'name': 'volume-9d006806-6309-415f-b72a-5e327ec77248', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '036fae78-2af0-49f1-85fa-245c29f35ebc', 'attached_at': '', 'detached_at': '', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'serial': '9d006806-6309-415f-b72a-5e327ec77248'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1119.296017] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3033a11-e7f8-4a9b-a4b2-fc499bdb6a86 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.316214] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b18d57-d406-440f-bc0d-751af265fd2c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.323148] env[61985]: WARNING nova.virt.vmwareapi.driver [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1119.323464] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1119.324194] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989db506-7247-4281-bb88-5378ae566c95 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.330715] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1119.330929] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-94fdcdf1-a0e2-4488-8590-fffefcc61ee0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.428065] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1119.428184] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1119.428474] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleting the datastore file [datastore1] 036fae78-2af0-49f1-85fa-245c29f35ebc {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1119.428923] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3dfa1b1-5a87-42f2-bcd4-eb4976d372c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.432551] env[61985]: DEBUG nova.scheduler.client.report [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1119.443496] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1119.443496] env[61985]: value = "task-936584" [ 1119.443496] env[61985]: _type = "Task" [ 1119.443496] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.456449] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936584, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.506440] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52db3acd-2dcc-91d1-92c8-1cc956e5817c, 'name': SearchDatastore_Task, 'duration_secs': 0.018374} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.511183] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.511552] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1119.511864] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.520192] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5219582b-89a0-44dc-1a63-40ba6744a442, 'name': SearchDatastore_Task, 'duration_secs': 0.008942} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.521713] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10f130ca-2520-4c0a-97a5-12c3a88535e3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.528674] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1119.528674] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a1bbce-ec1c-1b65-2531-af1709ea9619" [ 1119.528674] env[61985]: _type = "Task" [ 1119.528674] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.539227] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a1bbce-ec1c-1b65-2531-af1709ea9619, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.634020] env[61985]: DEBUG oslo_vmware.api [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Task: {'id': task-936581, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187838} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.634239] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.634426] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1119.634608] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1119.634815] env[61985]: INFO nova.compute.manager [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1119.635085] env[61985]: DEBUG oslo.service.loopingcall [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.635286] env[61985]: DEBUG nova.compute.manager [-] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1119.635382] env[61985]: DEBUG nova.network.neutron [-] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1119.711091] env[61985]: INFO nova.compute.manager [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Swapping old allocation on dict_keys(['aed7e5b3-c662-4538-8447-c4f67b460215']) held by migration ad0c29e8-fc04-41ba-a6ea-343ca3e29a1a for instance [ 1119.736161] env[61985]: DEBUG nova.scheduler.client.report [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Overwriting current allocation {'allocations': {'aed7e5b3-c662-4538-8447-c4f67b460215': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 144}}, 'project_id': '697ed07c609f4e1f86d317675a4749a8', 'user_id': '2f8df9858fb34d4cb63c2d735eb73319', 'consumer_generation': 1} on consumer 885a36d9-24c6-407a-9d6b-01cd97e90674 {{(pid=61985) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1119.830038] env[61985]: DEBUG oslo_concurrency.lockutils [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.830038] env[61985]: DEBUG oslo_concurrency.lockutils [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.830038] env[61985]: DEBUG nova.network.neutron [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1119.920069] env[61985]: DEBUG nova.compute.manager [req-24d18f87-d939-43b7-85e7-1cf84d07a86c req-ba65cb62-f922-4b45-a861-d910d4cb92ea service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Received event network-vif-deleted-ecf04033-48b6-4b81-ab17-57a89dc0be60 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1119.921102] env[61985]: INFO nova.compute.manager [req-24d18f87-d939-43b7-85e7-1cf84d07a86c req-ba65cb62-f922-4b45-a861-d910d4cb92ea service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Neutron deleted interface ecf04033-48b6-4b81-ab17-57a89dc0be60; detaching it from the instance and deleting it from the info cache [ 1119.921102] env[61985]: DEBUG nova.network.neutron [req-24d18f87-d939-43b7-85e7-1cf84d07a86c req-ba65cb62-f922-4b45-a861-d910d4cb92ea service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.940063] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.765s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.942775] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.703s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.942775] env[61985]: DEBUG nova.objects.instance [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lazy-loading 'resources' on Instance uuid f45760fa-2298-422e-84ff-83f99b19627d {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.954126] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936584, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.345416} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.954420] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.954590] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1119.954780] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1119.962672] env[61985]: INFO nova.scheduler.client.report [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Deleted allocations for instance 93dd0394-d7e1-42d5-a58f-fcd621c044b3 [ 1120.040068] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a1bbce-ec1c-1b65-2531-af1709ea9619, 'name': SearchDatastore_Task, 'duration_secs': 0.029779} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.040068] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.040068] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f/c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1120.040364] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.040364] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1120.040556] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea5e59d9-c283-41dc-87c2-a4b14f3d4b92 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.042588] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dc1fbd3-8ab8-45b6-b226-dc1ace5659cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.049875] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1120.049875] env[61985]: value = "task-936585" [ 1120.049875] env[61985]: _type = "Task" [ 1120.049875] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.053984] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1120.054193] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1120.055157] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17816e37-1d63-43fa-bfec-fd6295032d55 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.060105] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936585, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.063032] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1120.063032] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525bbd82-7c91-0f86-f86c-abe0c1954bf2" [ 1120.063032] env[61985]: _type = "Task" [ 1120.063032] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.071176] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525bbd82-7c91-0f86-f86c-abe0c1954bf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.398875] env[61985]: DEBUG nova.network.neutron [-] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.423187] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23e9751e-7c5e-47e4-93d4-7003bed7060f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.435776] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16747f3-046f-442f-b447-52655a7eed3d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.460643] env[61985]: INFO nova.virt.block_device [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Booting with volume 9d006806-6309-415f-b72a-5e327ec77248 at /dev/sdb [ 1120.485864] env[61985]: DEBUG nova.compute.manager [req-24d18f87-d939-43b7-85e7-1cf84d07a86c req-ba65cb62-f922-4b45-a861-d910d4cb92ea service nova] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Detach interface failed, port_id=ecf04033-48b6-4b81-ab17-57a89dc0be60, reason: Instance ef330828-c218-4b80-8b1a-6813bdb67c4d could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1120.487077] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4891f594-b4ab-4668-815b-5c2d00f3c1aa tempest-ServersTestMultiNic-703857195 tempest-ServersTestMultiNic-703857195-project-member] Lock "93dd0394-d7e1-42d5-a58f-fcd621c044b3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.058s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.517460] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3263bd1d-efa9-4ec4-b8e2-3e3ad93a7bd7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.531349] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda14294-b170-4d76-b281-a7b9579c2f1d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.573612] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-49073580-7369-4355-9c68-5830f408278d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.575866] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936585, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456009} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.578780] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f/c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1120.579020] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1120.580037] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b85ffb1c-70f5-4cfe-bd30-e67a57f0d48a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.585602] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]525bbd82-7c91-0f86-f86c-abe0c1954bf2, 'name': SearchDatastore_Task, 'duration_secs': 0.009187} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.588092] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1120.588092] env[61985]: value = "task-936587" [ 1120.588092] env[61985]: _type = "Task" [ 1120.588092] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.588566] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f6dbbea-2753-4f4b-8fb7-ef5aeeaf5e9a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.593108] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15b0243-7143-4e28-a492-d52198e2b7f4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.615366] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.616387] env[61985]: DEBUG nova.network.neutron [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance_info_cache with network_info: [{"id": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "address": "fa:16:3e:e1:d0:1f", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa64c525f-b3", "ovs_interfaceid": "a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.618977] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1120.618977] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527c270c-c569-bf57-cb11-c098a044ec7d" [ 1120.618977] env[61985]: _type = "Task" [ 1120.618977] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.634409] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55ea81f-3be7-4afd-8baf-c7e1dbf01345 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.640371] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527c270c-c569-bf57-cb11-c098a044ec7d, 'name': SearchDatastore_Task, 'duration_secs': 0.009478} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.640991] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.641278] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] f2bd7642-e20c-408b-9a7f-73428b89328a/f2bd7642-e20c-408b-9a7f-73428b89328a.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1120.641528] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c2fd798-db55-4b70-92ef-27ea8dc477f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.649253] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e752a58-fbe4-45f5-b2d0-7c5c7a13aa81 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.653781] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1120.653781] env[61985]: value = "task-936588" [ 1120.653781] env[61985]: _type = "Task" [ 1120.653781] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.668771] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936588, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.669140] env[61985]: DEBUG nova.virt.block_device [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Updating existing volume attachment record: 43788a7d-c57f-4cfc-835a-f62d960247e9 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1120.691593] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01984c81-ec0a-4dac-8321-b9f169a04dfd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.699968] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2c3c6a-0df4-4784-a9f4-8a27330e21b2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.731147] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10cd267a-002e-4b16-b49f-416cc10229ec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.739545] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cb61f2-720b-4a62-9db7-5d82fceb0033 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.756691] env[61985]: DEBUG nova.compute.provider_tree [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.902613] env[61985]: INFO nova.compute.manager [-] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Took 1.27 seconds to deallocate network for instance. [ 1121.100590] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070518} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.100880] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1121.101752] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96022f9f-77bc-41c2-80eb-84a77089ff5a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.124147] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f/c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1121.124749] env[61985]: DEBUG oslo_concurrency.lockutils [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-885a36d9-24c6-407a-9d6b-01cd97e90674" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.125142] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70bf9c40-300b-43a1-9770-52b13bedd0db {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.139929] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec97d2da-ce6b-4ccc-a870-ca50eae549c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.146683] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7120488c-53b4-4e58-bca0-6bf9bb078632 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.150159] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1121.150159] env[61985]: value = "task-936589" [ 1121.150159] env[61985]: _type = "Task" [ 1121.150159] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.160191] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936589, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.168977] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936588, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.430501} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.169282] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] f2bd7642-e20c-408b-9a7f-73428b89328a/f2bd7642-e20c-408b-9a7f-73428b89328a.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1121.169503] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1121.169752] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-469feb9f-6e14-43f3-8b48-25f01ac70357 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.176802] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1121.176802] env[61985]: value = "task-936590" [ 1121.176802] env[61985]: _type = "Task" [ 1121.176802] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.185930] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936590, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.260154] env[61985]: DEBUG nova.scheduler.client.report [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.410211] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.661896] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936589, 'name': ReconfigVM_Task, 'duration_secs': 0.296369} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.662324] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Reconfigured VM instance instance-0000006a to attach disk [datastore1] c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f/c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1121.663016] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0d00763-7d81-4c4e-b675-7efee8434235 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.670343] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1121.670343] env[61985]: value = "task-936591" [ 1121.670343] env[61985]: _type = "Task" [ 1121.670343] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.679108] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936591, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.686771] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936590, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067155} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.687877] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1121.687877] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181de835-49d5-4a3f-a4a0-0cb116e636bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.708654] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] f2bd7642-e20c-408b-9a7f-73428b89328a/f2bd7642-e20c-408b-9a7f-73428b89328a.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1121.708958] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98a8e8e1-89b3-47d3-9c27-bec696607d9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.734130] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1121.734130] env[61985]: value = "task-936592" [ 1121.734130] env[61985]: _type = "Task" [ 1121.734130] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.743509] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936592, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.765852] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.823s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.768666] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.359s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.768946] env[61985]: DEBUG nova.objects.instance [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lazy-loading 'resources' on Instance uuid ef330828-c218-4b80-8b1a-6813bdb67c4d {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1121.798799] env[61985]: INFO nova.scheduler.client.report [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted allocations for instance f45760fa-2298-422e-84ff-83f99b19627d [ 1122.181014] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936591, 'name': Rename_Task, 'duration_secs': 0.156536} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.181330] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1122.181581] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d627a75c-22f4-41a2-96d9-c0b8f6af5ec1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.188492] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1122.188492] env[61985]: value = "task-936593" [ 1122.188492] env[61985]: _type = "Task" [ 1122.188492] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.196890] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936593, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.245388] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936592, 'name': ReconfigVM_Task, 'duration_secs': 0.278025} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.245771] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Reconfigured VM instance instance-0000006b to attach disk [datastore1] f2bd7642-e20c-408b-9a7f-73428b89328a/f2bd7642-e20c-408b-9a7f-73428b89328a.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1122.246460] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-75f69756-99b0-4fdf-80b4-9773ca8044b6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.254179] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1122.254179] env[61985]: value = "task-936594" [ 1122.254179] env[61985]: _type = "Task" [ 1122.254179] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.263980] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936594, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.312191] env[61985]: DEBUG oslo_concurrency.lockutils [None req-06393b9b-2172-4cb1-aacb-2bbf456b5a41 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "f45760fa-2298-422e-84ff-83f99b19627d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.323s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.368214] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1122.368768] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bcb31fb8-73f0-4c9e-9f4b-ce316fe6a636 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.378696] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1122.378696] env[61985]: value = "task-936595" [ 1122.378696] env[61985]: _type = "Task" [ 1122.378696] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.391405] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936595, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.439370] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d7aac3-d672-4f97-bd72-8f2e63ddfb75 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.448406] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f08dade0-06cd-4e52-9c43-073cc93be699 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.483187] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c237d614-4024-4762-bb51-06179de2ce8f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.491327] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78f55c9-582a-40c0-b3d3-66f0bb31cb5c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.508327] env[61985]: DEBUG nova.compute.provider_tree [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.702449] env[61985]: DEBUG oslo_vmware.api [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936593, 'name': PowerOnVM_Task, 'duration_secs': 0.501196} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.702740] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1122.702942] env[61985]: INFO nova.compute.manager [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Took 7.40 seconds to spawn the instance on the hypervisor. [ 1122.703277] env[61985]: DEBUG nova.compute.manager [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1122.704229] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3e5da6-90c2-44fd-b3b4-f4e998130fa4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.733813] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1122.734272] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211486', 'volume_id': '03676853-7534-4379-bea5-3bf20b6efccd', 'name': 'volume-03676853-7534-4379-bea5-3bf20b6efccd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9079e605-e865-4eb1-9857-03769e4b5ba6', 'attached_at': '', 'detached_at': '', 'volume_id': '03676853-7534-4379-bea5-3bf20b6efccd', 'serial': '03676853-7534-4379-bea5-3bf20b6efccd'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1122.735453] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdf0a52-0fec-4f26-b932-7dfa92fb6a46 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.763883] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d15f9b5-5d97-4e68-9f93-94f9ba81449b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.774454] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936594, 'name': Rename_Task, 'duration_secs': 0.15923} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.787502] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1122.795900] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-03676853-7534-4379-bea5-3bf20b6efccd/volume-03676853-7534-4379-bea5-3bf20b6efccd.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1122.797416] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c4ac0d8-d953-4f29-8e7a-0e585fff5603 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.798658] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c69221af-684a-44df-9c51-6cac082b931e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.824092] env[61985]: DEBUG oslo_vmware.api [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1122.824092] env[61985]: value = "task-936596" [ 1122.824092] env[61985]: _type = "Task" [ 1122.824092] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.825788] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1122.825788] env[61985]: value = "task-936597" [ 1122.825788] env[61985]: _type = "Task" [ 1122.825788] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.837521] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.843030] env[61985]: DEBUG oslo_vmware.api [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936596, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.869616] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1122.869761] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1122.869909] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.870212] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1122.870394] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.871062] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1122.871329] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1122.871637] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1122.871855] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1122.872050] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1122.872245] env[61985]: DEBUG nova.virt.hardware [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.873151] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422abd00-2065-4b14-b86e-be56f82dc614 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.887280] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adcfd43-f8b2-4d7c-a766-a4459b5a9933 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.896536] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936595, 'name': PowerOffVM_Task, 'duration_secs': 0.268455} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.906028] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1122.906256] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1122.906515] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1122.906665] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.906846] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1122.907044] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.907202] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1122.907455] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1122.907947] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1122.907947] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1122.908106] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1122.908331] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.913710] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:ab:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52d657fa-a084-485b-858a-368bb739f4fc', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1122.921217] env[61985]: DEBUG oslo.service.loopingcall [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1122.921870] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc6bfb09-ccd0-4644-9924-05ddb3f58ffc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.932390] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1122.932997] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-87793fcf-25ba-465e-81fd-11a721af1c9a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.952846] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1122.952846] env[61985]: value = "task-936598" [ 1122.952846] env[61985]: _type = "Task" [ 1122.952846] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.955143] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1122.955143] env[61985]: value = "task-936599" [ 1122.955143] env[61985]: _type = "Task" [ 1122.955143] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.969114] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936599, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.970705] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936598, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.012052] env[61985]: DEBUG nova.scheduler.client.report [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1123.229069] env[61985]: INFO nova.compute.manager [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Took 16.11 seconds to build instance. [ 1123.350026] env[61985]: DEBUG oslo_vmware.api [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936596, 'name': ReconfigVM_Task, 'duration_secs': 0.411202} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.350026] env[61985]: DEBUG oslo_vmware.api [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936597, 'name': PowerOnVM_Task, 'duration_secs': 0.465714} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.350026] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-03676853-7534-4379-bea5-3bf20b6efccd/volume-03676853-7534-4379-bea5-3bf20b6efccd.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1123.354968] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1123.355242] env[61985]: INFO nova.compute.manager [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Took 5.52 seconds to spawn the instance on the hypervisor. [ 1123.355455] env[61985]: DEBUG nova.compute.manager [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1123.355830] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db18fbc7-fef4-4283-91cd-142b0364cfd8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.368239] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0803c28f-afd5-4ede-a727-07f7245aa712 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.381271] env[61985]: DEBUG oslo_vmware.api [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1123.381271] env[61985]: value = "task-936600" [ 1123.381271] env[61985]: _type = "Task" [ 1123.381271] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.395023] env[61985]: DEBUG oslo_vmware.api [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936600, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.472040] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936598, 'name': CreateVM_Task, 'duration_secs': 0.489206} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.475382] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1123.475783] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936599, 'name': ReconfigVM_Task, 'duration_secs': 0.166793} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.476603] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.476770] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.477121] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1123.477851] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67cd68d5-bbc7-4c7c-9366-885d0ece2666 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.480596] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341fbc32-b88d-455d-8aa1-ac95a4d45318 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.490466] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1123.490466] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52694c36-0ee6-bc07-d4a3-9ece02cc7a67" [ 1123.490466] env[61985]: _type = "Task" [ 1123.490466] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.512497] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1123.512778] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1123.512948] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.513164] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1123.513318] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.513820] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1123.514067] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1123.514242] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1123.514475] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1123.515141] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1123.515458] env[61985]: DEBUG nova.virt.hardware [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.520038] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a9efad8-4c23-4e24-94be-85a4845d5855 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.523253] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.755s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.530309] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1123.530309] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5294a42a-11c2-ba85-e017-b43db10a1a65" [ 1123.530309] env[61985]: _type = "Task" [ 1123.530309] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.534237] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52694c36-0ee6-bc07-d4a3-9ece02cc7a67, 'name': SearchDatastore_Task, 'duration_secs': 0.028866} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.537487] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.538250] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1123.538250] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.538250] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.538450] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1123.539185] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1187e7b-cad2-4bb8-acda-5607af8f6d2e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.548090] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5294a42a-11c2-ba85-e017-b43db10a1a65, 'name': SearchDatastore_Task, 'duration_secs': 0.01211} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.555765] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfiguring VM instance instance-0000005a to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1123.556284] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1123.556606] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1123.560416] env[61985]: INFO nova.scheduler.client.report [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Deleted allocations for instance ef330828-c218-4b80-8b1a-6813bdb67c4d [ 1123.565701] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c84434c-d621-4ff5-b2ad-6d9067e2bd1f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.581560] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4ca10b7-fd8d-4b80-8ba4-501dfd885f86 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.587771] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "ad01c86c-4752-4189-a361-afec6fd56495" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.588178] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "ad01c86c-4752-4189-a361-afec6fd56495" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.595128] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1123.595128] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b5b4c3-adcb-d823-9864-8f5a49baee26" [ 1123.595128] env[61985]: _type = "Task" [ 1123.595128] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.600738] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1123.600738] env[61985]: value = "task-936601" [ 1123.600738] env[61985]: _type = "Task" [ 1123.600738] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.609983] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b5b4c3-adcb-d823-9864-8f5a49baee26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.617981] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936601, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.731294] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b8e22b8e-fa0a-41fe-b0c7-2153fb3b771c tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.622s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.896975] env[61985]: INFO nova.compute.manager [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Took 10.41 seconds to build instance. [ 1123.901588] env[61985]: DEBUG oslo_vmware.api [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936600, 'name': ReconfigVM_Task, 'duration_secs': 0.305384} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.901912] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211486', 'volume_id': '03676853-7534-4379-bea5-3bf20b6efccd', 'name': 'volume-03676853-7534-4379-bea5-3bf20b6efccd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9079e605-e865-4eb1-9857-03769e4b5ba6', 'attached_at': '', 'detached_at': '', 'volume_id': '03676853-7534-4379-bea5-3bf20b6efccd', 'serial': '03676853-7534-4379-bea5-3bf20b6efccd'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1124.070644] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4f80e164-fa94-4dfb-bee8-03baeb17bbed tempest-ServerPasswordTestJSON-1512551341 tempest-ServerPasswordTestJSON-1512551341-project-member] Lock "ef330828-c218-4b80-8b1a-6813bdb67c4d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.563s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.094016] env[61985]: DEBUG nova.compute.manager [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1124.112566] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b5b4c3-adcb-d823-9864-8f5a49baee26, 'name': SearchDatastore_Task, 'duration_secs': 0.027538} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.117023] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-667f99d3-89d9-4421-a5c6-4258326d1369 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.120509] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936601, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.126310] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1124.126310] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d0a3a1-26b3-d062-ab73-0ce7369548f9" [ 1124.126310] env[61985]: _type = "Task" [ 1124.126310] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.133456] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d0a3a1-26b3-d062-ab73-0ce7369548f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.398504] env[61985]: DEBUG oslo_concurrency.lockutils [None req-41bf3193-6e52-4369-b8b3-bc34eea147e2 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "f2bd7642-e20c-408b-9a7f-73428b89328a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.919s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.612392] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936601, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.613508] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.613834] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.615323] env[61985]: INFO nova.compute.claims [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.634603] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d0a3a1-26b3-d062-ab73-0ce7369548f9, 'name': SearchDatastore_Task, 'duration_secs': 0.022783} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.634813] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.635130] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 036fae78-2af0-49f1-85fa-245c29f35ebc/036fae78-2af0-49f1-85fa-245c29f35ebc.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1124.635407] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4619e887-786e-4d26-9a65-f5d1a2250ba2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.643249] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1124.643249] env[61985]: value = "task-936602" [ 1124.643249] env[61985]: _type = "Task" [ 1124.643249] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.652081] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936602, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.729826] env[61985]: INFO nova.compute.manager [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Rebuilding instance [ 1124.774816] env[61985]: DEBUG nova.compute.manager [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1124.775788] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9fbf3e0-7ef9-4af6-8e93-67d797b72eaf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.937781] env[61985]: DEBUG nova.objects.instance [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lazy-loading 'flavor' on Instance uuid 9079e605-e865-4eb1-9857-03769e4b5ba6 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.119929] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936601, 'name': ReconfigVM_Task, 'duration_secs': 1.189488} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.123430] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfigured VM instance instance-0000005a to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1125.128075] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d797bd-f292-43f8-824d-7611caeb1919 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.167434] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 885a36d9-24c6-407a-9d6b-01cd97e90674/885a36d9-24c6-407a-9d6b-01cd97e90674.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1125.173918] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b268ce4-3847-4726-b852-8da96b1f327f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.197304] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936602, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.199542] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1125.199542] env[61985]: value = "task-936603" [ 1125.199542] env[61985]: _type = "Task" [ 1125.199542] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.212153] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936603, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.287968] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1125.288316] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-810db517-a573-402d-8ed4-1671998b7635 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.298443] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1125.298443] env[61985]: value = "task-936604" [ 1125.298443] env[61985]: _type = "Task" [ 1125.298443] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.311036] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936604, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.327679] env[61985]: DEBUG nova.compute.manager [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Stashing vm_state: active {{(pid=61985) _prep_resize /opt/stack/nova/nova/compute/manager.py:5630}} [ 1125.443878] env[61985]: DEBUG oslo_concurrency.lockutils [None req-30dc3ab9-eb78-4379-bf8d-cad050b5af00 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.342s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.674031] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936602, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602571} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.674681] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 036fae78-2af0-49f1-85fa-245c29f35ebc/036fae78-2af0-49f1-85fa-245c29f35ebc.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1125.674963] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1125.675382] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b1b064e-7e19-4aed-88e1-2da5f965de19 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.691042] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1125.691042] env[61985]: value = "task-936605" [ 1125.691042] env[61985]: _type = "Task" [ 1125.691042] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.708983] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936605, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.719460] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936603, 'name': ReconfigVM_Task, 'duration_secs': 0.319058} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.719460] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 885a36d9-24c6-407a-9d6b-01cd97e90674/885a36d9-24c6-407a-9d6b-01cd97e90674.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1125.720266] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe55d68-636b-44a6-890c-8d530223c773 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.748934] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925e1695-41ba-447b-9884-d39200428f8c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.776736] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba995f2-0601-4d97-a843-d60250e4a9e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.811245] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c2df8e-7b37-412f-8ad2-c26a01748bc2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.822839] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936604, 'name': PowerOffVM_Task, 'duration_secs': 0.115032} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.823117] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1125.827499] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1125.827499] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1125.827499] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08fea6db-f50f-4741-a450-f9e8b9c02ffa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.828624] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c48debe-ccf9-4d36-81fa-6b1a71a04af0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.837922] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1125.837922] env[61985]: value = "task-936606" [ 1125.837922] env[61985]: _type = "Task" [ 1125.837922] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.840265] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1125.843425] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-882f8cb1-da7d-4b7b-b003-7885af415b91 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.848821] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.854438] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936606, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.860329] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd31ee6-d15c-467d-aa2d-398440d65223 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.870434] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2985b86d-fc27-49db-b05a-777437d5e38d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.876780] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1125.877018] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1125.877241] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Deleting the datastore file [datastore1] f2bd7642-e20c-408b-9a7f-73428b89328a {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1125.877517] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0370c7c-f230-410f-93ec-9a9aaf1c5f48 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.908511] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b839dcfb-d9f3-4d32-a16a-44eac7b381a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.914577] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1125.914577] env[61985]: value = "task-936608" [ 1125.914577] env[61985]: _type = "Task" [ 1125.914577] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.923019] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6241a85a-17e5-4666-a0a7-db790a10e63e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.937040] env[61985]: DEBUG nova.compute.provider_tree [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.941527] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936608, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.188522] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "9079e605-e865-4eb1-9857-03769e4b5ba6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.188880] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.202327] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936605, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067899} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.202604] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1126.203420] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef98a045-5c76-4610-9513-2b5dbb429477 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.227015] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 036fae78-2af0-49f1-85fa-245c29f35ebc/036fae78-2af0-49f1-85fa-245c29f35ebc.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1126.227789] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ccc8790-9584-4e6a-90ee-f123a366ce2b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.248686] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1126.248686] env[61985]: value = "task-936609" [ 1126.248686] env[61985]: _type = "Task" [ 1126.248686] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.263333] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936609, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.351160] env[61985]: DEBUG oslo_vmware.api [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936606, 'name': PowerOnVM_Task, 'duration_secs': 0.446918} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.351460] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1126.427497] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936608, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102226} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.427813] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1126.428026] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1126.428229] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1126.442898] env[61985]: DEBUG nova.scheduler.client.report [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.692609] env[61985]: INFO nova.compute.manager [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Detaching volume 03676853-7534-4379-bea5-3bf20b6efccd [ 1126.740447] env[61985]: INFO nova.virt.block_device [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Attempting to driver detach volume 03676853-7534-4379-bea5-3bf20b6efccd from mountpoint /dev/sdb [ 1126.740824] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1126.741129] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211486', 'volume_id': '03676853-7534-4379-bea5-3bf20b6efccd', 'name': 'volume-03676853-7534-4379-bea5-3bf20b6efccd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9079e605-e865-4eb1-9857-03769e4b5ba6', 'attached_at': '', 'detached_at': '', 'volume_id': '03676853-7534-4379-bea5-3bf20b6efccd', 'serial': '03676853-7534-4379-bea5-3bf20b6efccd'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1126.742767] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d428cc-3401-4cbe-9ef5-05c94af23953 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.782969] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3600e4ee-bb0e-4930-b8e6-f767a491a750 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.790325] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936609, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.796125] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa196e81-d346-41ed-a226-bb69afd52148 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.818494] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efc2732-2481-4b90-a5a9-93e8af242ca3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.834734] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] The volume has not been displaced from its original location: [datastore1] volume-03676853-7534-4379-bea5-3bf20b6efccd/volume-03676853-7534-4379-bea5-3bf20b6efccd.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1126.840344] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1126.840684] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4bb97a4-c8d1-4ab1-bbbb-dd1d917c94f3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.861108] env[61985]: DEBUG oslo_vmware.api [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1126.861108] env[61985]: value = "task-936610" [ 1126.861108] env[61985]: _type = "Task" [ 1126.861108] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.872114] env[61985]: DEBUG oslo_vmware.api [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936610, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.948234] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.334s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.948818] env[61985]: DEBUG nova.compute.manager [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1126.951497] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.103s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.260126] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936609, 'name': ReconfigVM_Task, 'duration_secs': 0.904034} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.260451] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 036fae78-2af0-49f1-85fa-245c29f35ebc/036fae78-2af0-49f1-85fa-245c29f35ebc.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1127.261920] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_type': 'disk', 'encryption_secret_uuid': None, 'encryption_options': None, 'encryption_format': None, 'disk_bus': None, 'device_name': '/dev/sda', 'boot_index': 0, 'encrypted': False, 'size': 0, 'guest_format': None, 'image_id': 'fe6d817b-6194-440f-988a-f4a94c580922'}], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211478', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'name': 'volume-9d006806-6309-415f-b72a-5e327ec77248', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '036fae78-2af0-49f1-85fa-245c29f35ebc', 'attached_at': '', 'detached_at': '', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'serial': '9d006806-6309-415f-b72a-5e327ec77248'}, 'device_type': None, 'attachment_id': '43788a7d-c57f-4cfc-835a-f62d960247e9', 'mount_device': '/dev/sdb', 'disk_bus': None, 'delete_on_termination': False, 'boot_index': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=61985) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1127.262223] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1127.262480] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211478', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'name': 'volume-9d006806-6309-415f-b72a-5e327ec77248', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '036fae78-2af0-49f1-85fa-245c29f35ebc', 'attached_at': '', 'detached_at': '', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'serial': '9d006806-6309-415f-b72a-5e327ec77248'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1127.263315] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efded184-8d85-494e-ad45-0028c9c90a31 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.279216] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff1af6f-71be-478c-bbd5-5217f6ff99bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.303445] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-9d006806-6309-415f-b72a-5e327ec77248/volume-9d006806-6309-415f-b72a-5e327ec77248.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.303732] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04348f42-137e-48f8-883a-dfca4e012e57 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.322672] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1127.322672] env[61985]: value = "task-936611" [ 1127.322672] env[61985]: _type = "Task" [ 1127.322672] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.334889] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936611, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.374906] env[61985]: DEBUG oslo_vmware.api [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936610, 'name': ReconfigVM_Task, 'duration_secs': 0.312179} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.375345] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1127.380292] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce8fde5f-4f13-4c6c-b8d6-243b3a9815ae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.398407] env[61985]: DEBUG oslo_vmware.api [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1127.398407] env[61985]: value = "task-936612" [ 1127.398407] env[61985]: _type = "Task" [ 1127.398407] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.409317] env[61985]: DEBUG oslo_vmware.api [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936612, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.410534] env[61985]: INFO nova.compute.manager [None req-26dc8010-308c-4d98-bd69-4262fb942848 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance to original state: 'active' [ 1127.454381] env[61985]: DEBUG nova.compute.utils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1127.458486] env[61985]: INFO nova.compute.claims [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1127.461864] env[61985]: DEBUG nova.compute.manager [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1127.461972] env[61985]: DEBUG nova.network.neutron [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1127.473342] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1127.473593] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1127.473760] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1127.473955] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1127.474122] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1127.474280] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1127.474494] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1127.474660] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1127.474833] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1127.475009] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1127.475210] env[61985]: DEBUG nova.virt.hardware [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1127.476356] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5b49ab-cc2d-4815-a88c-a145194da274 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.485479] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5c2e01-bdf7-424c-823c-b37af4dd563f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.500931] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1127.506612] env[61985]: DEBUG oslo.service.loopingcall [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.508174] env[61985]: DEBUG nova.policy [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e23fdb33638143409613d9d4f26f199c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe531ec9f60548adb457351e19f8d51b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1127.509650] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1127.509894] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31ccb146-c5c6-40bb-af3f-32e631d08b4c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.529031] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1127.529031] env[61985]: value = "task-936613" [ 1127.529031] env[61985]: _type = "Task" [ 1127.529031] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.537296] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936613, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.813463] env[61985]: DEBUG nova.network.neutron [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Successfully created port: 98f24ba0-3320-41fe-b872-8d1df1571cb7 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1127.833756] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936611, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.908993] env[61985]: DEBUG oslo_vmware.api [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936612, 'name': ReconfigVM_Task, 'duration_secs': 0.162354} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.909329] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211486', 'volume_id': '03676853-7534-4379-bea5-3bf20b6efccd', 'name': 'volume-03676853-7534-4379-bea5-3bf20b6efccd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9079e605-e865-4eb1-9857-03769e4b5ba6', 'attached_at': '', 'detached_at': '', 'volume_id': '03676853-7534-4379-bea5-3bf20b6efccd', 'serial': '03676853-7534-4379-bea5-3bf20b6efccd'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1127.965730] env[61985]: DEBUG nova.compute.manager [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1127.970416] env[61985]: INFO nova.compute.resource_tracker [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating resource usage from migration b0648906-0c75-4305-9d80-b664e288823a [ 1128.042122] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936613, 'name': CreateVM_Task, 'duration_secs': 0.27021} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.042305] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1128.042730] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.042903] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.043276] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1128.043539] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ddd7383-df14-4f60-8663-104d19467f7d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.050911] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1128.050911] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522cde4b-f102-43ed-1b36-82b676fbd5ec" [ 1128.050911] env[61985]: _type = "Task" [ 1128.050911] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.063837] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522cde4b-f102-43ed-1b36-82b676fbd5ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.144771] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46de7403-1b2c-433f-b24a-6b58da98bb77 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.152802] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df46094f-7264-4811-9231-6fba36312971 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.183027] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645e964f-4884-4d9e-b008-09074a0fee22 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.190982] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dae7661-fbd2-49fd-8a4d-c7dd70f6d331 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.204489] env[61985]: DEBUG nova.compute.provider_tree [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.334899] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936611, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.460776] env[61985]: DEBUG nova.objects.instance [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lazy-loading 'flavor' on Instance uuid 9079e605-e865-4eb1-9857-03769e4b5ba6 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.563888] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522cde4b-f102-43ed-1b36-82b676fbd5ec, 'name': SearchDatastore_Task, 'duration_secs': 0.011207} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.564198] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.564467] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1128.564709] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.564862] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.565071] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.565366] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c9b456e0-0b7e-4914-a44a-f6bcc12c392b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.575085] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.575306] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1128.576078] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7cab9ec-6265-48be-a6fb-d8454ab2111d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.582107] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1128.582107] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5223c56c-2b2d-4f37-657c-0098faf635aa" [ 1128.582107] env[61985]: _type = "Task" [ 1128.582107] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.590376] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5223c56c-2b2d-4f37-657c-0098faf635aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.710024] env[61985]: DEBUG nova.scheduler.client.report [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.835967] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936611, 'name': ReconfigVM_Task, 'duration_secs': 1.451769} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.837385] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-9d006806-6309-415f-b72a-5e327ec77248/volume-9d006806-6309-415f-b72a-5e327ec77248.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.844676] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09f7aa67-6a8c-43a8-b82f-b6bdcc165039 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.861727] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1128.861727] env[61985]: value = "task-936614" [ 1128.861727] env[61985]: _type = "Task" [ 1128.861727] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.870872] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936614, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.979584] env[61985]: DEBUG nova.compute.manager [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1129.009916] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1129.010255] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1129.010427] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1129.010620] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1129.010778] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1129.010938] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1129.011170] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1129.011340] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1129.011519] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1129.011690] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1129.011874] env[61985]: DEBUG nova.virt.hardware [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1129.012772] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4613f2e5-cac6-49c6-8156-94199d2a8c5a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.022144] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a012d03-14bb-466e-8ad2-ebf89bf7e53a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.093576] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5223c56c-2b2d-4f37-657c-0098faf635aa, 'name': SearchDatastore_Task, 'duration_secs': 0.009509} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.094450] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92f946a8-93dc-4c6a-8895-06ac016ae79c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.100522] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1129.100522] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528fc356-8a62-9f0b-0cea-f9d2dc2e31c0" [ 1129.100522] env[61985]: _type = "Task" [ 1129.100522] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.108818] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528fc356-8a62-9f0b-0cea-f9d2dc2e31c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.209332] env[61985]: DEBUG nova.compute.manager [req-9eea7df0-82e3-4f43-aa74-5e856e092d51 req-b62a336d-e9d4-46b8-b940-ab1bf3b2c3b0 service nova] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Received event network-vif-plugged-98f24ba0-3320-41fe-b872-8d1df1571cb7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1129.209582] env[61985]: DEBUG oslo_concurrency.lockutils [req-9eea7df0-82e3-4f43-aa74-5e856e092d51 req-b62a336d-e9d4-46b8-b940-ab1bf3b2c3b0 service nova] Acquiring lock "ad01c86c-4752-4189-a361-afec6fd56495-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.209791] env[61985]: DEBUG oslo_concurrency.lockutils [req-9eea7df0-82e3-4f43-aa74-5e856e092d51 req-b62a336d-e9d4-46b8-b940-ab1bf3b2c3b0 service nova] Lock "ad01c86c-4752-4189-a361-afec6fd56495-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.209981] env[61985]: DEBUG oslo_concurrency.lockutils [req-9eea7df0-82e3-4f43-aa74-5e856e092d51 req-b62a336d-e9d4-46b8-b940-ab1bf3b2c3b0 service nova] Lock "ad01c86c-4752-4189-a361-afec6fd56495-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.210189] env[61985]: DEBUG nova.compute.manager [req-9eea7df0-82e3-4f43-aa74-5e856e092d51 req-b62a336d-e9d4-46b8-b940-ab1bf3b2c3b0 service nova] [instance: ad01c86c-4752-4189-a361-afec6fd56495] No waiting events found dispatching network-vif-plugged-98f24ba0-3320-41fe-b872-8d1df1571cb7 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1129.210365] env[61985]: WARNING nova.compute.manager [req-9eea7df0-82e3-4f43-aa74-5e856e092d51 req-b62a336d-e9d4-46b8-b940-ab1bf3b2c3b0 service nova] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Received unexpected event network-vif-plugged-98f24ba0-3320-41fe-b872-8d1df1571cb7 for instance with vm_state building and task_state spawning. [ 1129.212486] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.261s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.212667] env[61985]: INFO nova.compute.manager [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Migrating [ 1129.356337] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "885a36d9-24c6-407a-9d6b-01cd97e90674" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.356642] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.356872] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "885a36d9-24c6-407a-9d6b-01cd97e90674-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.357084] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.357270] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.359562] env[61985]: INFO nova.compute.manager [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Terminating instance [ 1129.361631] env[61985]: DEBUG nova.compute.manager [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1129.361798] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1129.362092] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c75ce32-9c27-4b37-926a-dc3f67f2f75b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.375698] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936614, 'name': ReconfigVM_Task, 'duration_secs': 0.395187} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.377106] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211478', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'name': 'volume-9d006806-6309-415f-b72a-5e327ec77248', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '036fae78-2af0-49f1-85fa-245c29f35ebc', 'attached_at': '', 'detached_at': '', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'serial': '9d006806-6309-415f-b72a-5e327ec77248'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1129.377804] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1129.377804] env[61985]: value = "task-936615" [ 1129.377804] env[61985]: _type = "Task" [ 1129.377804] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.378016] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38e6daeb-d46a-4a4d-b1a9-d1865f8a8baa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.388604] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.390651] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1129.390651] env[61985]: value = "task-936616" [ 1129.390651] env[61985]: _type = "Task" [ 1129.390651] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.400957] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936616, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.472746] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4ebd1706-b519-40d6-88e1-ac6c8b3448fc tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.284s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.611542] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528fc356-8a62-9f0b-0cea-f9d2dc2e31c0, 'name': SearchDatastore_Task, 'duration_secs': 0.01191} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.612014] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.612259] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] f2bd7642-e20c-408b-9a7f-73428b89328a/f2bd7642-e20c-408b-9a7f-73428b89328a.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1129.612571] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-862be82a-fe8a-45bb-9ece-6f5fa01560b1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.620087] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1129.620087] env[61985]: value = "task-936617" [ 1129.620087] env[61985]: _type = "Task" [ 1129.620087] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.628110] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936617, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.727288] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.727618] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.727761] env[61985]: DEBUG nova.network.neutron [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1129.786896] env[61985]: DEBUG nova.network.neutron [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Successfully updated port: 98f24ba0-3320-41fe-b872-8d1df1571cb7 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1129.818564] env[61985]: DEBUG nova.compute.manager [req-c0b54cb3-fb13-46f4-8600-e396fc852d18 req-a5603a1a-248a-4c61-87d2-541d0d84d4c7 service nova] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Received event network-changed-98f24ba0-3320-41fe-b872-8d1df1571cb7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1129.818850] env[61985]: DEBUG nova.compute.manager [req-c0b54cb3-fb13-46f4-8600-e396fc852d18 req-a5603a1a-248a-4c61-87d2-541d0d84d4c7 service nova] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Refreshing instance network info cache due to event network-changed-98f24ba0-3320-41fe-b872-8d1df1571cb7. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1129.819598] env[61985]: DEBUG oslo_concurrency.lockutils [req-c0b54cb3-fb13-46f4-8600-e396fc852d18 req-a5603a1a-248a-4c61-87d2-541d0d84d4c7 service nova] Acquiring lock "refresh_cache-ad01c86c-4752-4189-a361-afec6fd56495" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.819771] env[61985]: DEBUG oslo_concurrency.lockutils [req-c0b54cb3-fb13-46f4-8600-e396fc852d18 req-a5603a1a-248a-4c61-87d2-541d0d84d4c7 service nova] Acquired lock "refresh_cache-ad01c86c-4752-4189-a361-afec6fd56495" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.819958] env[61985]: DEBUG nova.network.neutron [req-c0b54cb3-fb13-46f4-8600-e396fc852d18 req-a5603a1a-248a-4c61-87d2-541d0d84d4c7 service nova] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Refreshing network info cache for port 98f24ba0-3320-41fe-b872-8d1df1571cb7 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1129.890603] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936615, 'name': PowerOffVM_Task, 'duration_secs': 0.197538} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.890957] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1129.891272] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1129.891520] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211465', 'volume_id': '9954307d-32c0-4c02-bf62-bbfe223b3c98', 'name': 'volume-9954307d-32c0-4c02-bf62-bbfe223b3c98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '885a36d9-24c6-407a-9d6b-01cd97e90674', 'attached_at': '2024-09-18T01:33:33.000000', 'detached_at': '', 'volume_id': '9954307d-32c0-4c02-bf62-bbfe223b3c98', 'serial': '9954307d-32c0-4c02-bf62-bbfe223b3c98'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1129.892503] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e3fc6a-9cdd-4cca-9965-bbac15682cf0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.906304] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936616, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.923798] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3bb43d-179f-44e3-bd9d-f9f047de86b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.933732] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba5d298-97ad-4f70-b2cc-cf21b8eb8a28 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.964158] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a194f2f-71c3-4ad3-a178-bb40e771041b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.991630] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] The volume has not been displaced from its original location: [datastore2] volume-9954307d-32c0-4c02-bf62-bbfe223b3c98/volume-9954307d-32c0-4c02-bf62-bbfe223b3c98.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1129.999804] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfiguring VM instance instance-0000005a to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1130.001872] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dca92f11-5349-4dac-ab90-2936bb8f3439 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.028411] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1130.028411] env[61985]: value = "task-936618" [ 1130.028411] env[61985]: _type = "Task" [ 1130.028411] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.039770] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936618, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.130443] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936617, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448754} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.130722] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] f2bd7642-e20c-408b-9a7f-73428b89328a/f2bd7642-e20c-408b-9a7f-73428b89328a.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1130.130945] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1130.131229] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f9956161-ea53-4f4f-8e1f-395d10e0f015 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.138206] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1130.138206] env[61985]: value = "task-936619" [ 1130.138206] env[61985]: _type = "Task" [ 1130.138206] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.147722] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936619, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.290737] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "refresh_cache-ad01c86c-4752-4189-a361-afec6fd56495" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.366567] env[61985]: DEBUG nova.network.neutron [req-c0b54cb3-fb13-46f4-8600-e396fc852d18 req-a5603a1a-248a-4c61-87d2-541d0d84d4c7 service nova] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1130.402339] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936616, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.497561] env[61985]: DEBUG nova.network.neutron [req-c0b54cb3-fb13-46f4-8600-e396fc852d18 req-a5603a1a-248a-4c61-87d2-541d0d84d4c7 service nova] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.513552] env[61985]: DEBUG nova.network.neutron [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance_info_cache with network_info: [{"id": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "address": "fa:16:3e:d8:5d:8b", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7303cfe7-ec", "ovs_interfaceid": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.520784] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "9079e605-e865-4eb1-9857-03769e4b5ba6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.521078] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.521288] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "9079e605-e865-4eb1-9857-03769e4b5ba6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.521462] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.521726] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.523978] env[61985]: INFO nova.compute.manager [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Terminating instance [ 1130.525978] env[61985]: DEBUG nova.compute.manager [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1130.526212] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1130.527362] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4010a64-4841-4739-98b0-fa86113d45b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.539095] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1130.541964] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3020a51b-295f-48b3-846d-faf1963a90be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.543548] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936618, 'name': ReconfigVM_Task, 'duration_secs': 0.213582} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.544068] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Reconfigured VM instance instance-0000005a to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1130.549066] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7507e69d-c8dd-45fe-9e5a-cb4b956ac7c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.560737] env[61985]: DEBUG oslo_vmware.api [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1130.560737] env[61985]: value = "task-936620" [ 1130.560737] env[61985]: _type = "Task" [ 1130.560737] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.566732] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1130.566732] env[61985]: value = "task-936621" [ 1130.566732] env[61985]: _type = "Task" [ 1130.566732] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.573441] env[61985]: DEBUG oslo_vmware.api [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936620, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.579596] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936621, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.649041] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936619, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068605} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.649394] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1130.650293] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99dfc2bf-eaa5-4a83-8126-13791baa76dc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.676278] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] f2bd7642-e20c-408b-9a7f-73428b89328a/f2bd7642-e20c-408b-9a7f-73428b89328a.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1130.676726] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-904e764b-560b-4a13-b439-ec6bf8821ac9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.700406] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1130.700406] env[61985]: value = "task-936622" [ 1130.700406] env[61985]: _type = "Task" [ 1130.700406] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.710417] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936622, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.902662] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936616, 'name': Rename_Task, 'duration_secs': 1.19349} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.903018] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1130.903241] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb6041b6-5660-4444-84b0-ae77e67f7da4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.910582] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1130.910582] env[61985]: value = "task-936623" [ 1130.910582] env[61985]: _type = "Task" [ 1130.910582] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.918780] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936623, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.000665] env[61985]: DEBUG oslo_concurrency.lockutils [req-c0b54cb3-fb13-46f4-8600-e396fc852d18 req-a5603a1a-248a-4c61-87d2-541d0d84d4c7 service nova] Releasing lock "refresh_cache-ad01c86c-4752-4189-a361-afec6fd56495" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.001097] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "refresh_cache-ad01c86c-4752-4189-a361-afec6fd56495" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.001263] env[61985]: DEBUG nova.network.neutron [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1131.016149] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.073063] env[61985]: DEBUG oslo_vmware.api [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936620, 'name': PowerOffVM_Task, 'duration_secs': 0.23304} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.073920] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1131.074141] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1131.074383] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b898d64-04d9-4c17-a1ae-e464632a4439 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.078407] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936621, 'name': ReconfigVM_Task, 'duration_secs': 0.204568} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.078969] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211465', 'volume_id': '9954307d-32c0-4c02-bf62-bbfe223b3c98', 'name': 'volume-9954307d-32c0-4c02-bf62-bbfe223b3c98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '885a36d9-24c6-407a-9d6b-01cd97e90674', 'attached_at': '2024-09-18T01:33:33.000000', 'detached_at': '', 'volume_id': '9954307d-32c0-4c02-bf62-bbfe223b3c98', 'serial': '9954307d-32c0-4c02-bf62-bbfe223b3c98'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1131.079288] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1131.079994] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73985d3-6258-497b-a4ea-43d49d82f7c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.086863] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1131.087112] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4c5bdf58-4425-4b8a-a557-2e4e1ff90b4f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.161324] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1131.161619] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1131.161859] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleting the datastore file [datastore1] 885a36d9-24c6-407a-9d6b-01cd97e90674 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1131.162190] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb72fcea-d7ae-4da2-a7e3-9b2075cb6f2e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.170452] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1131.170452] env[61985]: value = "task-936626" [ 1131.170452] env[61985]: _type = "Task" [ 1131.170452] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.179554] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936626, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.210752] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936622, 'name': ReconfigVM_Task, 'duration_secs': 0.260751} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.211091] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Reconfigured VM instance instance-0000006b to attach disk [datastore1] f2bd7642-e20c-408b-9a7f-73428b89328a/f2bd7642-e20c-408b-9a7f-73428b89328a.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1131.211720] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7531618d-9ea2-47de-9d8e-3fcbe79f0c61 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.219558] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1131.219558] env[61985]: value = "task-936627" [ 1131.219558] env[61985]: _type = "Task" [ 1131.219558] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.227841] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936627, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.421761] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936623, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.533534] env[61985]: DEBUG nova.network.neutron [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1131.555171] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1131.555440] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1131.555586] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleting the datastore file [datastore2] 9079e605-e865-4eb1-9857-03769e4b5ba6 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1131.556215] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2de0987-dcb8-4991-b9fd-b47c31516e54 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.564046] env[61985]: DEBUG oslo_vmware.api [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1131.564046] env[61985]: value = "task-936628" [ 1131.564046] env[61985]: _type = "Task" [ 1131.564046] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.572239] env[61985]: DEBUG oslo_vmware.api [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.682434] env[61985]: DEBUG oslo_vmware.api [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936626, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175906} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.682799] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1131.683062] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1131.683267] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1131.683489] env[61985]: INFO nova.compute.manager [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Took 2.32 seconds to destroy the instance on the hypervisor. [ 1131.683691] env[61985]: DEBUG oslo.service.loopingcall [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1131.683917] env[61985]: DEBUG nova.compute.manager [-] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1131.684095] env[61985]: DEBUG nova.network.neutron [-] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1131.731753] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936627, 'name': Rename_Task, 'duration_secs': 0.156734} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.732061] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1131.732321] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cda69cad-be93-43f1-a7fa-29e2a1a2ea81 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.740600] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1131.740600] env[61985]: value = "task-936629" [ 1131.740600] env[61985]: _type = "Task" [ 1131.740600] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.749192] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.780836] env[61985]: DEBUG nova.network.neutron [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Updating instance_info_cache with network_info: [{"id": "98f24ba0-3320-41fe-b872-8d1df1571cb7", "address": "fa:16:3e:7c:96:f1", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98f24ba0-33", "ovs_interfaceid": "98f24ba0-3320-41fe-b872-8d1df1571cb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.922150] env[61985]: DEBUG oslo_vmware.api [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936623, 'name': PowerOnVM_Task, 'duration_secs': 0.562749} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.922510] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1131.922786] env[61985]: DEBUG nova.compute.manager [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1131.923548] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fae0bda-b94e-47a8-b60e-f5a9196aabb6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.076118] env[61985]: DEBUG oslo_vmware.api [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.202605} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.076277] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1132.076551] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1132.076761] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1132.077034] env[61985]: INFO nova.compute.manager [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Took 1.55 seconds to destroy the instance on the hypervisor. [ 1132.077326] env[61985]: DEBUG oslo.service.loopingcall [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1132.077590] env[61985]: DEBUG nova.compute.manager [-] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1132.077590] env[61985]: DEBUG nova.network.neutron [-] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1132.263607] env[61985]: DEBUG oslo_vmware.api [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936629, 'name': PowerOnVM_Task, 'duration_secs': 0.45621} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.264019] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1132.264341] env[61985]: DEBUG nova.compute.manager [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1132.265520] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbc884e-3d05-4759-83a3-df7c5c17cef7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.284908] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "refresh_cache-ad01c86c-4752-4189-a361-afec6fd56495" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.284908] env[61985]: DEBUG nova.compute.manager [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Instance network_info: |[{"id": "98f24ba0-3320-41fe-b872-8d1df1571cb7", "address": "fa:16:3e:7c:96:f1", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98f24ba0-33", "ovs_interfaceid": "98f24ba0-3320-41fe-b872-8d1df1571cb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1132.285431] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:96:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55520f67-d092-4eb7-940f-d7cceaa1ca1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '98f24ba0-3320-41fe-b872-8d1df1571cb7', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1132.294337] env[61985]: DEBUG oslo.service.loopingcall [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1132.294613] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1132.294844] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7320bcda-90be-41e9-9f9e-4cddb532debc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.316422] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1132.316422] env[61985]: value = "task-936630" [ 1132.316422] env[61985]: _type = "Task" [ 1132.316422] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.325959] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936630, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.378437] env[61985]: DEBUG nova.compute.manager [req-41fac55f-0e65-4ea2-aa3f-e6dc0ccc36af req-735cded1-5984-464b-adc7-a71c29a5ffb1 service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Received event network-vif-deleted-a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1132.378687] env[61985]: INFO nova.compute.manager [req-41fac55f-0e65-4ea2-aa3f-e6dc0ccc36af req-735cded1-5984-464b-adc7-a71c29a5ffb1 service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Neutron deleted interface a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7; detaching it from the instance and deleting it from the info cache [ 1132.378896] env[61985]: DEBUG nova.network.neutron [req-41fac55f-0e65-4ea2-aa3f-e6dc0ccc36af req-735cded1-5984-464b-adc7-a71c29a5ffb1 service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.448208] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.448208] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.448590] env[61985]: DEBUG nova.objects.instance [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61985) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1132.530327] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c620d8-34af-432a-8f0a-e5d2c3d040aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.554023] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance 'c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f' progress to 0 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1132.791550] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.822680] env[61985]: DEBUG nova.network.neutron [-] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.833087] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936630, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.853281] env[61985]: DEBUG oslo_concurrency.lockutils [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.853684] env[61985]: DEBUG oslo_concurrency.lockutils [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.854187] env[61985]: INFO nova.compute.manager [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Shelving [ 1132.883382] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f0d547d3-7a25-4be2-ab73-2586f7962a42 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.894262] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84072dcf-37f5-4263-9da9-6535f31f69c8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.931645] env[61985]: DEBUG nova.compute.manager [req-41fac55f-0e65-4ea2-aa3f-e6dc0ccc36af req-735cded1-5984-464b-adc7-a71c29a5ffb1 service nova] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Detach interface failed, port_id=a64c525f-b3b2-4f27-82d3-f0ac5df1f5c7, reason: Instance 885a36d9-24c6-407a-9d6b-01cd97e90674 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1133.064129] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1133.064464] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53c77758-f304-4e57-8ff3-051f6bf6a52e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.073578] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1133.073578] env[61985]: value = "task-936631" [ 1133.073578] env[61985]: _type = "Task" [ 1133.073578] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.083671] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936631, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.173676] env[61985]: DEBUG nova.network.neutron [-] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.328288] env[61985]: INFO nova.compute.manager [-] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Took 1.64 seconds to deallocate network for instance. [ 1133.328659] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936630, 'name': CreateVM_Task, 'duration_secs': 0.63536} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.330698] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1133.331570] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.331747] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.332092] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1133.332359] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db3988a4-d0fc-45f6-8123-4b5dc045bbac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.338551] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1133.338551] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5238474c-ec56-a4aa-0419-a43da62e02a0" [ 1133.338551] env[61985]: _type = "Task" [ 1133.338551] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.347726] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5238474c-ec56-a4aa-0419-a43da62e02a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.362041] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1133.362041] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-adea60f4-121c-43ee-b8f6-489bc834afd0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.371424] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1133.371424] env[61985]: value = "task-936632" [ 1133.371424] env[61985]: _type = "Task" [ 1133.371424] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.383051] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936632, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.459744] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fed10e6b-2c5f-492f-a9d9-c309fbb7c052 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.460804] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.670s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.460922] env[61985]: DEBUG nova.objects.instance [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61985) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1133.512585] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "f2bd7642-e20c-408b-9a7f-73428b89328a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.512834] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "f2bd7642-e20c-408b-9a7f-73428b89328a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.513071] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "f2bd7642-e20c-408b-9a7f-73428b89328a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.513271] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "f2bd7642-e20c-408b-9a7f-73428b89328a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.513451] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "f2bd7642-e20c-408b-9a7f-73428b89328a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.515838] env[61985]: INFO nova.compute.manager [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Terminating instance [ 1133.517434] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "refresh_cache-f2bd7642-e20c-408b-9a7f-73428b89328a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.517601] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquired lock "refresh_cache-f2bd7642-e20c-408b-9a7f-73428b89328a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.517819] env[61985]: DEBUG nova.network.neutron [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1133.584520] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936631, 'name': PowerOffVM_Task, 'duration_secs': 0.428766} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.584802] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1133.585012] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance 'c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f' progress to 17 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1133.677973] env[61985]: INFO nova.compute.manager [-] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Took 1.60 seconds to deallocate network for instance. [ 1133.809662] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "036fae78-2af0-49f1-85fa-245c29f35ebc" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.809930] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.849208] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5238474c-ec56-a4aa-0419-a43da62e02a0, 'name': SearchDatastore_Task, 'duration_secs': 0.015536} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.849553] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.849799] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1133.850094] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.850265] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.850484] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1133.850779] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c3a7f31c-aea4-49c1-8b46-ea533182f56f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.862929] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1133.863143] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1133.863885] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98ae2f8e-a430-4bdf-bbc1-001166e74b4c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.870257] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1133.870257] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526d22d9-2c07-2921-b4cd-d379d5b9c0c9" [ 1133.870257] env[61985]: _type = "Task" [ 1133.870257] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.874674] env[61985]: INFO nova.compute.manager [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Took 0.55 seconds to detach 1 volumes for instance. [ 1133.886533] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526d22d9-2c07-2921-b4cd-d379d5b9c0c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.891336] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936632, 'name': PowerOffVM_Task, 'duration_secs': 0.274141} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.891423] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1133.892257] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff1d9c8-695d-4ee9-adce-dcd192b247f9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.913628] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0507de-1097-44cd-a2da-56547acbe6bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.036569] env[61985]: DEBUG nova.network.neutron [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1134.088513] env[61985]: DEBUG nova.network.neutron [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.091671] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1134.091966] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1134.092157] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1134.092373] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1134.092533] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1134.092704] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1134.092953] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1134.093164] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1134.093347] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1134.093545] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1134.093743] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1134.099835] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3cab20c1-8f06-4111-a7bb-f2660005d3ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.117673] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1134.117673] env[61985]: value = "task-936633" [ 1134.117673] env[61985]: _type = "Task" [ 1134.117673] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.127450] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936633, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.185021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.313166] env[61985]: INFO nova.compute.manager [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Detaching volume 9d006806-6309-415f-b72a-5e327ec77248 [ 1134.348592] env[61985]: INFO nova.virt.block_device [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Attempting to driver detach volume 9d006806-6309-415f-b72a-5e327ec77248 from mountpoint /dev/sdb [ 1134.348893] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1134.349126] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211478', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'name': 'volume-9d006806-6309-415f-b72a-5e327ec77248', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '036fae78-2af0-49f1-85fa-245c29f35ebc', 'attached_at': '', 'detached_at': '', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'serial': '9d006806-6309-415f-b72a-5e327ec77248'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1134.350019] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d064837-2708-4dee-8075-63a92c3ce3a5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.375845] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6da364d-9d26-418a-badc-ee04e9512810 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.384443] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526d22d9-2c07-2921-b4cd-d379d5b9c0c9, 'name': SearchDatastore_Task, 'duration_secs': 0.014642} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.387028] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-195aaccd-606f-4ad6-9e65-fc94b24e61e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.389685] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2979fcf1-c349-41d6-9c9c-5603b544a4be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.392720] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.396310] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1134.396310] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5223861d-280c-b9a8-caf1-ddf91ccbf652" [ 1134.396310] env[61985]: _type = "Task" [ 1134.396310] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.416484] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d569740-4d43-49a6-a169-4c33e79022ae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.419699] env[61985]: DEBUG nova.compute.manager [req-26921a89-ae93-4912-8b96-fcfd4ed09344 req-e585e46b-28c7-44cf-84f4-7cdbbb6401ff service nova] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Received event network-vif-deleted-1a081dc2-f497-4708-a3af-0b373d04cb85 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1134.428670] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1134.428964] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5223861d-280c-b9a8-caf1-ddf91ccbf652, 'name': SearchDatastore_Task, 'duration_secs': 0.009277} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.441023] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b81ad731-b532-44ea-8e97-9d9b47a40d43 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.442449] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.442913] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] ad01c86c-4752-4189-a361-afec6fd56495/ad01c86c-4752-4189-a361-afec6fd56495.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1134.444086] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] The volume has not been displaced from its original location: [datastore2] volume-9d006806-6309-415f-b72a-5e327ec77248/volume-9d006806-6309-415f-b72a-5e327ec77248.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1134.449107] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1134.449460] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-847d1920-132b-4403-ad6c-d3a3477fec03 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.451818] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2d7e945-9790-4bfa-ab70-e7e766378b9a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.471893] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1134.471893] env[61985]: value = "task-936634" [ 1134.471893] env[61985]: _type = "Task" [ 1134.471893] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.473907] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9391dc95-16da-4c1e-844c-6926c8ae587f tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.475075] env[61985]: DEBUG oslo_vmware.api [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1134.475075] env[61985]: value = "task-936635" [ 1134.475075] env[61985]: _type = "Task" [ 1134.475075] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.479947] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.295s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.480249] env[61985]: DEBUG nova.objects.instance [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lazy-loading 'resources' on Instance uuid 9079e605-e865-4eb1-9857-03769e4b5ba6 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.482127] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1134.482127] env[61985]: value = "task-936636" [ 1134.482127] env[61985]: _type = "Task" [ 1134.482127] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.494523] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.494864] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936634, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.495434] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.495601] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1134.495755] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Rebuilding the list of instances to heal {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9936}} [ 1134.505216] env[61985]: DEBUG oslo_vmware.api [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936635, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.505611] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936636, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.543430] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.543585] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.600107] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Releasing lock "refresh_cache-f2bd7642-e20c-408b-9a7f-73428b89328a" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.600679] env[61985]: DEBUG nova.compute.manager [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1134.600986] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1134.602383] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-479cda4e-d669-40fd-b483-b7126eba12e2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.611520] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1134.611815] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f9b2207-5158-4ca1-83f1-7872ab375cad {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.620280] env[61985]: DEBUG oslo_vmware.api [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1134.620280] env[61985]: value = "task-936637" [ 1134.620280] env[61985]: _type = "Task" [ 1134.620280] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.634539] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936633, 'name': ReconfigVM_Task, 'duration_secs': 0.192148} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.637960] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance 'c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f' progress to 33 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1134.641675] env[61985]: DEBUG oslo_vmware.api [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936637, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.986994] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936634, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.002877] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936636, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499629} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.006757] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] ad01c86c-4752-4189-a361-afec6fd56495/ad01c86c-4752-4189-a361-afec6fd56495.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1135.007128] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1135.010127] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Skipping network cache update for instance because it is being deleted. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9949}} [ 1135.010305] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1135.012664] env[61985]: DEBUG oslo_vmware.api [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936635, 'name': ReconfigVM_Task, 'duration_secs': 0.254611} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.012992] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fd4d50a-6f65-4f09-a56e-9edd64e58005 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.017182] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1135.024528] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32871146-76ef-41dc-9343-ac68f101b697 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.041171] env[61985]: DEBUG oslo_vmware.api [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1135.041171] env[61985]: value = "task-936639" [ 1135.041171] env[61985]: _type = "Task" [ 1135.041171] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.042467] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1135.042467] env[61985]: value = "task-936638" [ 1135.042467] env[61985]: _type = "Task" [ 1135.042467] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.046318] env[61985]: DEBUG nova.compute.manager [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1135.064785] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936638, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.065510] env[61985]: DEBUG oslo_vmware.api [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936639, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.073646] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "refresh_cache-59e742d9-c51a-47dd-9ef2-2f070eae94c4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.073797] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquired lock "refresh_cache-59e742d9-c51a-47dd-9ef2-2f070eae94c4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.074220] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Forcefully refreshing network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1135.074220] env[61985]: DEBUG nova.objects.instance [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lazy-loading 'info_cache' on Instance uuid 59e742d9-c51a-47dd-9ef2-2f070eae94c4 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.138221] env[61985]: DEBUG oslo_vmware.api [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936637, 'name': PowerOffVM_Task, 'duration_secs': 0.370673} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.138545] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1135.138734] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1135.138995] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc45f38c-cc69-4329-9ddd-4499fa52fa8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.144934] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1135.145214] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1135.145434] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1135.145537] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1135.145710] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1135.145918] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1135.146163] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1135.146335] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1135.146509] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1135.146679] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1135.146947] env[61985]: DEBUG nova.virt.hardware [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1135.152359] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1135.155421] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20460743-c76b-49ae-9a5b-a7e9dfcc7d70 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.171094] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1135.171395] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1135.171525] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Deleting the datastore file [datastore1] f2bd7642-e20c-408b-9a7f-73428b89328a {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1135.172281] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8142fcc6-ae1a-477e-8f40-0851c3cbc1ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.178211] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1135.178211] env[61985]: value = "task-936641" [ 1135.178211] env[61985]: _type = "Task" [ 1135.178211] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.182908] env[61985]: DEBUG oslo_vmware.api [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for the task: (returnval){ [ 1135.182908] env[61985]: value = "task-936642" [ 1135.182908] env[61985]: _type = "Task" [ 1135.182908] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.192552] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936641, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.198644] env[61985]: DEBUG oslo_vmware.api [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936642, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.229514] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7951106a-89bf-492d-9c4b-ee034a6422e2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.238513] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4fe280-1677-453d-b645-7622c7776eb7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.269614] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf23b4e8-b19a-4caf-a39a-818640050d4f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.278436] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ebb42e-d623-4f90-90f2-3a3d06ce46ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.295136] env[61985]: DEBUG nova.compute.provider_tree [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.484834] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936634, 'name': CreateSnapshot_Task, 'duration_secs': 0.619415} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.485140] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1135.486052] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b808e9-7127-4bf0-a6d6-625c2e93081e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.561652] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936638, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.142154} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.561899] env[61985]: DEBUG oslo_vmware.api [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936639, 'name': ReconfigVM_Task, 'duration_secs': 0.194613} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.564349] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1135.564724] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211478', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'name': 'volume-9d006806-6309-415f-b72a-5e327ec77248', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '036fae78-2af0-49f1-85fa-245c29f35ebc', 'attached_at': '', 'detached_at': '', 'volume_id': '9d006806-6309-415f-b72a-5e327ec77248', 'serial': '9d006806-6309-415f-b72a-5e327ec77248'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1135.568472] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d1ad54-a988-4506-841a-b5a19ecda161 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.597608] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] ad01c86c-4752-4189-a361-afec6fd56495/ad01c86c-4752-4189-a361-afec6fd56495.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.599040] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.600081] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8e8a09e-8e5e-405d-944c-b8a6c88f4d3c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.621299] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1135.621299] env[61985]: value = "task-936643" [ 1135.621299] env[61985]: _type = "Task" [ 1135.621299] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.630717] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936643, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.693398] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936641, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.700020] env[61985]: DEBUG oslo_vmware.api [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Task: {'id': task-936642, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.439601} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.700298] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1135.700521] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1135.700728] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1135.700935] env[61985]: INFO nova.compute.manager [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1135.701223] env[61985]: DEBUG oslo.service.loopingcall [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1135.701422] env[61985]: DEBUG nova.compute.manager [-] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1135.701537] env[61985]: DEBUG nova.network.neutron [-] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1135.717754] env[61985]: DEBUG nova.network.neutron [-] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1135.802504] env[61985]: DEBUG nova.scheduler.client.report [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.006299] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1136.006590] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-eaf6e494-e554-4dee-92ba-d3e0bce0ac9b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.019436] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1136.019436] env[61985]: value = "task-936644" [ 1136.019436] env[61985]: _type = "Task" [ 1136.019436] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.029063] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936644, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.121681] env[61985]: DEBUG nova.objects.instance [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lazy-loading 'flavor' on Instance uuid 036fae78-2af0-49f1-85fa-245c29f35ebc {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1136.136409] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936643, 'name': ReconfigVM_Task, 'duration_secs': 0.266852} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.137557] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Reconfigured VM instance instance-0000006c to attach disk [datastore1] ad01c86c-4752-4189-a361-afec6fd56495/ad01c86c-4752-4189-a361-afec6fd56495.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1136.138248] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0a546031-6255-40fe-b559-10cdf2f18d83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.146461] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1136.146461] env[61985]: value = "task-936645" [ 1136.146461] env[61985]: _type = "Task" [ 1136.146461] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.164890] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936645, 'name': Rename_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.191853] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936641, 'name': ReconfigVM_Task, 'duration_secs': 0.561901} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.192172] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1136.192977] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882a0139-cc4e-4469-8d0f-42e18308609f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.217241] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f/c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1136.217978] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-819848e9-2d7d-4373-8ffe-bd2119cf8be5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.231071] env[61985]: DEBUG nova.network.neutron [-] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.239399] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1136.239399] env[61985]: value = "task-936646" [ 1136.239399] env[61985]: _type = "Task" [ 1136.239399] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.249354] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936646, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.308286] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.828s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.311038] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.918s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.311293] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.313408] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.715s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.314991] env[61985]: INFO nova.compute.claims [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1136.331917] env[61985]: INFO nova.scheduler.client.report [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleted allocations for instance 9079e605-e865-4eb1-9857-03769e4b5ba6 [ 1136.338706] env[61985]: INFO nova.scheduler.client.report [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleted allocations for instance 885a36d9-24c6-407a-9d6b-01cd97e90674 [ 1136.531037] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936644, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.657736] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936645, 'name': Rename_Task, 'duration_secs': 0.158035} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.658084] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1136.658325] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bca4183f-6e39-476c-a0b5-f166175a1bd2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.666894] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1136.666894] env[61985]: value = "task-936647" [ 1136.666894] env[61985]: _type = "Task" [ 1136.666894] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.678377] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936647, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.737880] env[61985]: INFO nova.compute.manager [-] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Took 1.04 seconds to deallocate network for instance. [ 1136.756600] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936646, 'name': ReconfigVM_Task, 'duration_secs': 0.41442} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.757148] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Reconfigured VM instance instance-0000006a to attach disk [datastore1] c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f/c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1136.757587] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance 'c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f' progress to 50 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1136.841761] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a84f8911-1617-4889-87c4-6450158b5d58 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "9079e605-e865-4eb1-9857-03769e4b5ba6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.320s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.846020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-aa1a3feb-0054-452f-b330-f7990b5c0269 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "885a36d9-24c6-407a-9d6b-01cd97e90674" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.489s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.914618] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Updating instance_info_cache with network_info: [{"id": "876d546b-a049-4c69-817f-aa6e7d7a2fc2", "address": "fa:16:3e:68:e0:60", "network": {"id": "1e26ca38-9fb1-446d-a361-f97f07381e0b", "bridge": "br-int", "label": "tempest-ServersTestJSON-455221638-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe531ec9f60548adb457351e19f8d51b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap876d546b-a0", "ovs_interfaceid": "876d546b-a049-4c69-817f-aa6e7d7a2fc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.032897] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936644, 'name': CloneVM_Task, 'duration_secs': 1.009537} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.034465] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Created linked-clone VM from snapshot [ 1137.034465] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f6a1e6-0969-41b2-b9f5-d30141205e9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.043089] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Uploading image 2bf9b33d-ad4a-42a8-8df6-48e63defc809 {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1137.069710] env[61985]: DEBUG oslo_vmware.rw_handles [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1137.069710] env[61985]: value = "vm-211491" [ 1137.069710] env[61985]: _type = "VirtualMachine" [ 1137.069710] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1137.070058] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a8e0d485-466c-4d56-b4a2-491e1cb794d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.078134] env[61985]: DEBUG oslo_vmware.rw_handles [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lease: (returnval){ [ 1137.078134] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d11a4d-018d-bca6-5a73-ac9de7ece9e7" [ 1137.078134] env[61985]: _type = "HttpNfcLease" [ 1137.078134] env[61985]: } obtained for exporting VM: (result){ [ 1137.078134] env[61985]: value = "vm-211491" [ 1137.078134] env[61985]: _type = "VirtualMachine" [ 1137.078134] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1137.078469] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the lease: (returnval){ [ 1137.078469] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d11a4d-018d-bca6-5a73-ac9de7ece9e7" [ 1137.078469] env[61985]: _type = "HttpNfcLease" [ 1137.078469] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1137.085763] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1137.085763] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d11a4d-018d-bca6-5a73-ac9de7ece9e7" [ 1137.085763] env[61985]: _type = "HttpNfcLease" [ 1137.085763] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1137.135732] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776477db-25ee-48fa-9845-4df5f72c793d tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.325s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.176943] env[61985]: DEBUG oslo_vmware.api [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936647, 'name': PowerOnVM_Task, 'duration_secs': 0.472541} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.178094] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1137.178094] env[61985]: INFO nova.compute.manager [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Took 8.20 seconds to spawn the instance on the hypervisor. [ 1137.178094] env[61985]: DEBUG nova.compute.manager [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1137.178493] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c97065-ffa7-46ac-a995-43675097e1b7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.249699] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.264629] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-314637b3-0b25-4d5d-bec2-14725e25607e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.292356] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7448bf74-310d-4e64-9277-9027b5410503 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.313850] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance 'c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f' progress to 67 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1137.421021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Releasing lock "refresh_cache-59e742d9-c51a-47dd-9ef2-2f070eae94c4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.421021] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Updated the network info_cache for instance {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10003}} [ 1137.421021] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.421021] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.421021] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.421021] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.422777] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.423262] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.424095] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 1137.424359] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.562322] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93b5e06-48cc-4083-92ce-9b9714bd7998 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.570858] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fed025f-609a-4996-8875-2b8425cf2e19 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.617548] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c94d46-1672-4721-a92d-dd24fa69f15f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.624535] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1137.624535] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d11a4d-018d-bca6-5a73-ac9de7ece9e7" [ 1137.624535] env[61985]: _type = "HttpNfcLease" [ 1137.624535] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1137.626666] env[61985]: DEBUG oslo_vmware.rw_handles [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1137.626666] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d11a4d-018d-bca6-5a73-ac9de7ece9e7" [ 1137.626666] env[61985]: _type = "HttpNfcLease" [ 1137.626666] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1137.627440] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a123f02d-b52a-4033-a23f-1207946643f7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.630700] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570e589d-b4e9-495a-8cc0-325245fe1178 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.646172] env[61985]: DEBUG nova.compute.provider_tree [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.649437] env[61985]: DEBUG oslo_vmware.rw_handles [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b80463-fd82-53fd-564a-3c86c0ea5af8/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1137.649620] env[61985]: DEBUG oslo_vmware.rw_handles [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b80463-fd82-53fd-564a-3c86c0ea5af8/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1137.719632] env[61985]: INFO nova.compute.manager [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Took 13.12 seconds to build instance. [ 1137.792022] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-bab210b9-2715-4466-b11a-00d6151fab3d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.886538] env[61985]: DEBUG nova.network.neutron [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Port 7303cfe7-ecb8-4ee1-9b69-414635f73c1b binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1137.927584] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.976672] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "45cb51fd-3876-4875-9a5d-b7760072ef55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.976933] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.152863] env[61985]: DEBUG nova.scheduler.client.report [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1138.220185] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8b7c0443-c1f0-4ebb-8510-2dd7cd20db98 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "ad01c86c-4752-4189-a361-afec6fd56495" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.632s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.220553] env[61985]: DEBUG oslo_concurrency.lockutils [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "036fae78-2af0-49f1-85fa-245c29f35ebc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.220780] env[61985]: DEBUG oslo_concurrency.lockutils [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.220989] env[61985]: DEBUG oslo_concurrency.lockutils [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "036fae78-2af0-49f1-85fa-245c29f35ebc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.221200] env[61985]: DEBUG oslo_concurrency.lockutils [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.221373] env[61985]: DEBUG oslo_concurrency.lockutils [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.223892] env[61985]: INFO nova.compute.manager [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Terminating instance [ 1138.225786] env[61985]: DEBUG nova.compute.manager [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1138.226014] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1138.227035] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd944ffa-64e4-459f-b50b-9a6bae381f4a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.236821] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1138.237089] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1daeea3e-23f0-4276-b358-a0cfe0849eaa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.244287] env[61985]: DEBUG oslo_vmware.api [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1138.244287] env[61985]: value = "task-936649" [ 1138.244287] env[61985]: _type = "Task" [ 1138.244287] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.254150] env[61985]: DEBUG oslo_vmware.api [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.289823] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "ad01c86c-4752-4189-a361-afec6fd56495" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.290108] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "ad01c86c-4752-4189-a361-afec6fd56495" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.290303] env[61985]: DEBUG nova.compute.manager [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1138.291507] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e00382-da9e-45b8-ae18-d43c4cb8aa39 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.299558] env[61985]: DEBUG nova.compute.manager [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61985) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3374}} [ 1138.300284] env[61985]: DEBUG nova.objects.instance [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lazy-loading 'flavor' on Instance uuid ad01c86c-4752-4189-a361-afec6fd56495 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.479989] env[61985]: DEBUG nova.compute.manager [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1138.658522] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.659272] env[61985]: DEBUG nova.compute.manager [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1138.662457] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.413s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.662763] env[61985]: DEBUG nova.objects.instance [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lazy-loading 'resources' on Instance uuid f2bd7642-e20c-408b-9a7f-73428b89328a {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1138.755251] env[61985]: DEBUG oslo_vmware.api [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936649, 'name': PowerOffVM_Task, 'duration_secs': 0.254236} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.755251] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1138.756179] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1138.758042] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f104e558-a90a-49f8-bed0-2dac399e77f4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.806553] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1138.806959] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ee36803-e5dd-4adf-81bc-276ab1e39807 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.816266] env[61985]: DEBUG oslo_vmware.api [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1138.816266] env[61985]: value = "task-936651" [ 1138.816266] env[61985]: _type = "Task" [ 1138.816266] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.828456] env[61985]: DEBUG oslo_vmware.api [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936651, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.830921] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1138.831332] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1138.831598] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleting the datastore file [datastore2] 036fae78-2af0-49f1-85fa-245c29f35ebc {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1138.831945] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bea1ad20-ba45-4569-8d21-70a30c598648 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.841284] env[61985]: DEBUG oslo_vmware.api [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1138.841284] env[61985]: value = "task-936652" [ 1138.841284] env[61985]: _type = "Task" [ 1138.841284] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.852320] env[61985]: DEBUG oslo_vmware.api [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936652, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.909857] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.910220] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.910412] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.006719] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.096140] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "41627143-d60c-4b18-8192-ace45be71d0d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.096477] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.168245] env[61985]: DEBUG nova.compute.utils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1139.170451] env[61985]: DEBUG nova.compute.manager [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1139.170574] env[61985]: DEBUG nova.network.neutron [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1139.332462] env[61985]: DEBUG oslo_vmware.api [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936651, 'name': PowerOffVM_Task, 'duration_secs': 0.231489} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.333125] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1139.333496] env[61985]: DEBUG nova.compute.manager [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1139.334682] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999dfcb2-b9aa-437c-88da-662cb61da134 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.338888] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4167d5-c523-4982-9ea7-b300a4cf8b41 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.355235] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7b476f-a7b9-4f72-a91f-9e410875f041 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.365652] env[61985]: DEBUG oslo_vmware.api [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936652, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220778} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.366762] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1139.367803] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1139.368277] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1139.369154] env[61985]: INFO nova.compute.manager [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1139.369154] env[61985]: DEBUG oslo.service.loopingcall [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1139.369154] env[61985]: DEBUG nova.compute.manager [-] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1139.369359] env[61985]: DEBUG nova.network.neutron [-] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1139.404572] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94d593c-e0d0-4646-8b32-6cf7ca90c486 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.406295] env[61985]: DEBUG nova.policy [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '905fa37e1f784a56b3edb9818b03d685', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23a4809e353a4ecaa1bd53612fb4afc5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1139.415575] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e040394b-bc0c-4518-96dc-3182dcc92bea {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.436939] env[61985]: DEBUG nova.compute.provider_tree [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.601762] env[61985]: DEBUG nova.compute.manager [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1139.674883] env[61985]: DEBUG nova.compute.manager [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1139.858043] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d0f4a98a-fbcd-4749-a2f6-0931200e4690 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "ad01c86c-4752-4189-a361-afec6fd56495" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.567s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.863168] env[61985]: DEBUG nova.network.neutron [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Successfully created port: 7f79fbde-374c-4af0-b37b-98c170b5a562 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1139.942169] env[61985]: DEBUG nova.scheduler.client.report [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1139.961525] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.961796] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.962810] env[61985]: DEBUG nova.network.neutron [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1140.130918] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.178546] env[61985]: DEBUG nova.compute.manager [req-f7fc0f96-f794-4346-a4a7-3b2ca9e982f8 req-735ce9c7-3044-4b38-8e66-64378840a264 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Received event network-vif-deleted-52d657fa-a084-485b-858a-368bb739f4fc {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1140.178838] env[61985]: INFO nova.compute.manager [req-f7fc0f96-f794-4346-a4a7-3b2ca9e982f8 req-735ce9c7-3044-4b38-8e66-64378840a264 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Neutron deleted interface 52d657fa-a084-485b-858a-368bb739f4fc; detaching it from the instance and deleting it from the info cache [ 1140.179067] env[61985]: DEBUG nova.network.neutron [req-f7fc0f96-f794-4346-a4a7-3b2ca9e982f8 req-735ce9c7-3044-4b38-8e66-64378840a264 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.448035] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.785s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.453586] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.523s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.453586] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.453586] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1140.453586] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.445s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.453586] env[61985]: INFO nova.compute.claims [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1140.457101] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe3cdbe-d43f-49df-9a50-5f541d8182e4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.469596] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473d5a2f-74cc-4e6c-98e3-67ccef8d1ba6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.477014] env[61985]: INFO nova.scheduler.client.report [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Deleted allocations for instance f2bd7642-e20c-408b-9a7f-73428b89328a [ 1140.496591] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911c4325-8519-4e2b-bed2-bea0125a7b7b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.506016] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962629e6-efc3-4f89-a9df-01bcc14f40d8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.542169] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180382MB free_disk=43GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1140.542336] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.652873] env[61985]: DEBUG nova.network.neutron [-] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.688985] env[61985]: DEBUG nova.compute.manager [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1140.691080] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e6cd266-d061-449c-be97-9dbdabea9446 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.701736] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ed9b4d-4d2b-4285-9c84-86c817f2170c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.721546] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1140.721811] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1140.721989] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1140.722200] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1140.722359] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1140.722890] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1140.722890] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1140.723025] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1140.723253] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1140.723332] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1140.723595] env[61985]: DEBUG nova.virt.hardware [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1140.724348] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae604d1-07bc-42d4-95e2-1acecc38aac4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.736660] env[61985]: DEBUG nova.compute.manager [req-f7fc0f96-f794-4346-a4a7-3b2ca9e982f8 req-735ce9c7-3044-4b38-8e66-64378840a264 service nova] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Detach interface failed, port_id=52d657fa-a084-485b-858a-368bb739f4fc, reason: Instance 036fae78-2af0-49f1-85fa-245c29f35ebc could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1140.740525] env[61985]: DEBUG nova.network.neutron [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance_info_cache with network_info: [{"id": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "address": "fa:16:3e:d8:5d:8b", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7303cfe7-ec", "ovs_interfaceid": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.745134] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67c212e-ded9-4d7c-b3c3-94b4fa0c258e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.986030] env[61985]: DEBUG oslo_concurrency.lockutils [None req-e89b5b00-c124-4e00-9a5e-8d2ecdab3480 tempest-ServerShowV254Test-2080889887 tempest-ServerShowV254Test-2080889887-project-member] Lock "f2bd7642-e20c-408b-9a7f-73428b89328a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.473s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.156174] env[61985]: INFO nova.compute.manager [-] [instance: 036fae78-2af0-49f1-85fa-245c29f35ebc] Took 1.79 seconds to deallocate network for instance. [ 1141.249296] env[61985]: DEBUG oslo_concurrency.lockutils [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.374242] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "ad01c86c-4752-4189-a361-afec6fd56495" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.374604] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "ad01c86c-4752-4189-a361-afec6fd56495" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.374832] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "ad01c86c-4752-4189-a361-afec6fd56495-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.374970] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "ad01c86c-4752-4189-a361-afec6fd56495-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.375266] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "ad01c86c-4752-4189-a361-afec6fd56495-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.377786] env[61985]: INFO nova.compute.manager [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Terminating instance [ 1141.380435] env[61985]: DEBUG nova.compute.manager [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1141.380717] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1141.381577] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558e54b2-e8b6-4e4c-a142-8452f9127b48 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.391648] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1141.391977] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-30ae1361-74a9-4861-a7be-326a544c13f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.537363] env[61985]: DEBUG nova.network.neutron [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Successfully updated port: 7f79fbde-374c-4af0-b37b-98c170b5a562 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1141.622627] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c66131-4a68-4892-bb27-e027fd2ef526 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.631544] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b05fbd-93ef-4822-8c9c-0c098bd27148 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.662698] env[61985]: DEBUG oslo_concurrency.lockutils [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.663731] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3fc403c-85b6-4467-84c4-a03a35a83696 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.673009] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62f4c49-0ce2-48eb-a22e-f78b73e0d660 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.688735] env[61985]: DEBUG nova.compute.provider_tree [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.772370] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a74f46-93c4-49a3-b68f-461e49bc7de8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.794088] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81b53bf-bde6-46a7-a07a-8594978fdcb8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.802801] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance 'c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f' progress to 83 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1142.041374] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.041549] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.041709] env[61985]: DEBUG nova.network.neutron [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1142.192504] env[61985]: DEBUG nova.scheduler.client.report [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1142.206738] env[61985]: DEBUG nova.compute.manager [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Received event network-vif-plugged-7f79fbde-374c-4af0-b37b-98c170b5a562 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1142.207089] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] Acquiring lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.207347] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] Lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.207541] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] Lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.207759] env[61985]: DEBUG nova.compute.manager [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] No waiting events found dispatching network-vif-plugged-7f79fbde-374c-4af0-b37b-98c170b5a562 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1142.207963] env[61985]: WARNING nova.compute.manager [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Received unexpected event network-vif-plugged-7f79fbde-374c-4af0-b37b-98c170b5a562 for instance with vm_state building and task_state spawning. [ 1142.208187] env[61985]: DEBUG nova.compute.manager [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Received event network-changed-7f79fbde-374c-4af0-b37b-98c170b5a562 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1142.208366] env[61985]: DEBUG nova.compute.manager [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Refreshing instance network info cache due to event network-changed-7f79fbde-374c-4af0-b37b-98c170b5a562. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1142.208585] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] Acquiring lock "refresh_cache-8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.267766] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1142.268131] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1142.268193] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleting the datastore file [datastore1] ad01c86c-4752-4189-a361-afec6fd56495 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1142.268464] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61005330-e946-413a-ae9b-4660d808af68 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.275617] env[61985]: DEBUG oslo_vmware.api [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1142.275617] env[61985]: value = "task-936654" [ 1142.275617] env[61985]: _type = "Task" [ 1142.275617] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.284349] env[61985]: DEBUG oslo_vmware.api [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936654, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.310633] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1142.310995] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-81afe67c-5561-4cef-b985-0078c42cb22a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.320557] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1142.320557] env[61985]: value = "task-936655" [ 1142.320557] env[61985]: _type = "Task" [ 1142.320557] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.329380] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936655, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.577729] env[61985]: DEBUG nova.network.neutron [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1142.697880] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.247s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.698446] env[61985]: DEBUG nova.compute.manager [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1142.701187] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.570s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.702623] env[61985]: INFO nova.compute.claims [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1142.721257] env[61985]: DEBUG nova.network.neutron [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Updating instance_info_cache with network_info: [{"id": "7f79fbde-374c-4af0-b37b-98c170b5a562", "address": "fa:16:3e:ac:ba:51", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f79fbde-37", "ovs_interfaceid": "7f79fbde-374c-4af0-b37b-98c170b5a562", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.786597] env[61985]: DEBUG oslo_vmware.api [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936654, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173485} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.787077] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1142.787172] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1142.787743] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1142.787743] env[61985]: INFO nova.compute.manager [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Took 1.41 seconds to destroy the instance on the hypervisor. [ 1142.787939] env[61985]: DEBUG oslo.service.loopingcall [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1142.788180] env[61985]: DEBUG nova.compute.manager [-] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1142.788294] env[61985]: DEBUG nova.network.neutron [-] [instance: ad01c86c-4752-4189-a361-afec6fd56495] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1142.830745] env[61985]: DEBUG oslo_vmware.api [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936655, 'name': PowerOnVM_Task, 'duration_secs': 0.460663} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.831043] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1142.831244] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-42ba47c2-6b12-42ab-b16d-2cbeef9a2fc6 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance 'c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f' progress to 100 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1143.206912] env[61985]: DEBUG nova.compute.utils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1143.210469] env[61985]: DEBUG nova.compute.manager [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1143.210652] env[61985]: DEBUG nova.network.neutron [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1143.223754] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.224143] env[61985]: DEBUG nova.compute.manager [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Instance network_info: |[{"id": "7f79fbde-374c-4af0-b37b-98c170b5a562", "address": "fa:16:3e:ac:ba:51", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f79fbde-37", "ovs_interfaceid": "7f79fbde-374c-4af0-b37b-98c170b5a562", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1143.224469] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] Acquired lock "refresh_cache-8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.224673] env[61985]: DEBUG nova.network.neutron [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Refreshing network info cache for port 7f79fbde-374c-4af0-b37b-98c170b5a562 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1143.225766] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:ba:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f79fbde-374c-4af0-b37b-98c170b5a562', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1143.233280] env[61985]: DEBUG oslo.service.loopingcall [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1143.236139] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1143.236614] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74f63b8d-522f-4714-97e6-2abd296c40a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.253339] env[61985]: DEBUG nova.policy [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f8df9858fb34d4cb63c2d735eb73319', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '697ed07c609f4e1f86d317675a4749a8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1143.258319] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1143.258319] env[61985]: value = "task-936656" [ 1143.258319] env[61985]: _type = "Task" [ 1143.258319] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.266716] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936656, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.550311] env[61985]: DEBUG nova.network.neutron [-] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.619783] env[61985]: DEBUG nova.network.neutron [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Updated VIF entry in instance network info cache for port 7f79fbde-374c-4af0-b37b-98c170b5a562. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1143.620274] env[61985]: DEBUG nova.network.neutron [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Updating instance_info_cache with network_info: [{"id": "7f79fbde-374c-4af0-b37b-98c170b5a562", "address": "fa:16:3e:ac:ba:51", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f79fbde-37", "ovs_interfaceid": "7f79fbde-374c-4af0-b37b-98c170b5a562", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.710254] env[61985]: DEBUG nova.network.neutron [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Successfully created port: eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1143.712027] env[61985]: DEBUG nova.compute.manager [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1143.772334] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936656, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.050066] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d837d8-cc89-40e9-bb2e-bb05841124ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.054025] env[61985]: INFO nova.compute.manager [-] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Took 1.27 seconds to deallocate network for instance. [ 1144.061215] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6cb917f-ea29-43a2-ae8c-2ed955b38eb2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.094035] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e6eef5-2ba5-4abf-9869-87dc53ef16db {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.103142] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a73e96-984b-4f15-9653-141126485411 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.117556] env[61985]: DEBUG nova.compute.provider_tree [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1144.123595] env[61985]: DEBUG oslo_concurrency.lockutils [req-8a2d340a-a0d4-4bd3-8f65-5096a438851e req-d8d513ab-392c-4d79-bb56-6c63e755d552 service nova] Releasing lock "refresh_cache-8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.269423] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936656, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.297559] env[61985]: DEBUG nova.compute.manager [req-150dc776-64d2-433d-acac-493bae9e36e7 req-816c86c0-c094-471e-956d-3730b420f3d5 service nova] [instance: ad01c86c-4752-4189-a361-afec6fd56495] Received event network-vif-deleted-98f24ba0-3320-41fe-b872-8d1df1571cb7 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1144.561894] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.620698] env[61985]: DEBUG nova.scheduler.client.report [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1144.728446] env[61985]: DEBUG nova.compute.manager [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1144.756755] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1144.757035] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1144.757211] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1144.757409] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1144.757564] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1144.757761] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1144.757997] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1144.758223] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1144.758423] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1144.758598] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1144.758819] env[61985]: DEBUG nova.virt.hardware [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1144.759845] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37da98c3-d39f-452c-843a-bbde39073056 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.775425] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6af9603-83d5-4283-8775-6b512c62dce1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.779883] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936656, 'name': CreateVM_Task, 'duration_secs': 1.50712} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.780108] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1144.781240] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.781441] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.781914] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1144.782227] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3dda4c73-44c0-49aa-a568-9821e2df9bae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.795919] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1144.795919] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5274262b-4931-84ee-c0ee-7be7f7d013e3" [ 1144.795919] env[61985]: _type = "Task" [ 1144.795919] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.804820] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5274262b-4931-84ee-c0ee-7be7f7d013e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.126014] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.126693] env[61985]: DEBUG nova.compute.manager [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1145.129528] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.587s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.258958] env[61985]: DEBUG nova.network.neutron [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Successfully updated port: eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1145.307466] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5274262b-4931-84ee-c0ee-7be7f7d013e3, 'name': SearchDatastore_Task, 'duration_secs': 0.013294} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.307923] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.308110] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1145.308353] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.308537] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.308808] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1145.309193] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b213cb48-b718-4987-a1b5-a8d329580a73 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.318481] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1145.318684] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1145.319463] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa8bc6a0-88f5-4356-9813-abdaadcdbf04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.325427] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1145.325427] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52821870-aa22-7832-6922-ea393f4c1a77" [ 1145.325427] env[61985]: _type = "Task" [ 1145.325427] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.333490] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52821870-aa22-7832-6922-ea393f4c1a77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.408294] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.408612] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.408853] env[61985]: DEBUG nova.compute.manager [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Going to confirm migration 7 {{(pid=61985) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4789}} [ 1145.633302] env[61985]: DEBUG nova.compute.utils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1145.640646] env[61985]: DEBUG nova.compute.manager [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1145.640900] env[61985]: DEBUG nova.network.neutron [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1145.681720] env[61985]: DEBUG nova.policy [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4f5b5ce76e8b4fae9c905bc7a39b683e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4923009773c047519b883745c59e3516', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1145.761996] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.762172] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.762540] env[61985]: DEBUG nova.network.neutron [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1145.837805] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52821870-aa22-7832-6922-ea393f4c1a77, 'name': SearchDatastore_Task, 'duration_secs': 0.013329} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.838713] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d3d90e4-4d46-4269-bfa2-29faaf9cfc66 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.845674] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1145.845674] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52038857-23bb-e43c-5d62-4298c363cefb" [ 1145.845674] env[61985]: _type = "Task" [ 1145.845674] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.855018] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52038857-23bb-e43c-5d62-4298c363cefb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.005184] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.005184] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquired lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.005184] env[61985]: DEBUG nova.network.neutron [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1146.005184] env[61985]: DEBUG nova.objects.instance [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lazy-loading 'info_cache' on Instance uuid c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.035793] env[61985]: DEBUG nova.network.neutron [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Successfully created port: bb5ccfcb-7752-47cd-b569-3c2861ea2a73 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1146.144744] env[61985]: DEBUG nova.compute.manager [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1146.148245] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Applying migration context for instance c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f as it has an incoming, in-progress migration b0648906-0c75-4305-9d80-b664e288823a. Migration status is finished {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1146.149400] env[61985]: INFO nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating resource usage from migration b0648906-0c75-4305-9d80-b664e288823a [ 1146.172431] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 59e742d9-c51a-47dd-9ef2-2f070eae94c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1146.172739] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 073a7668-39e6-480d-9350-835a0282b456 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1146.173033] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 44361f7b-4609-476f-b4a9-58a7851e6e92 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1146.173224] env[61985]: WARNING nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 036fae78-2af0-49f1-85fa-245c29f35ebc is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1146.173431] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1146.173640] env[61985]: WARNING nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance ad01c86c-4752-4189-a361-afec6fd56495 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1146.173904] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Migration b0648906-0c75-4305-9d80-b664e288823a is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1146.174052] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1146.174259] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1146.174463] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 45cb51fd-3876-4875-9a5d-b7760072ef55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1146.174786] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 41627143-d60c-4b18-8192-ace45be71d0d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1146.174963] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1146.175206] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1146.298470] env[61985]: DEBUG nova.network.neutron [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1146.334645] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c989698f-fc2f-4b19-9985-b858ee8afc6c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.349886] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527df2b3-cbf0-499c-8310-d1f63dbcc3f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.359369] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52038857-23bb-e43c-5d62-4298c363cefb, 'name': SearchDatastore_Task, 'duration_secs': 0.014084} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.384179] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.384492] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be/8b9791c2-bf5f-495a-8c5e-4e45ff4b91be.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1146.387531] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c960646-c8ea-43ab-abc1-7a06cf5faf62 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.389952] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a96873c6-e60a-45c9-975c-da295ef9710b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.400485] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2a8bea-32b8-4b35-859a-c91142487071 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.404710] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1146.404710] env[61985]: value = "task-936657" [ 1146.404710] env[61985]: _type = "Task" [ 1146.404710] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.415847] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1146.426774] env[61985]: DEBUG nova.compute.manager [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received event network-vif-plugged-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1146.426938] env[61985]: DEBUG oslo_concurrency.lockutils [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] Acquiring lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.427190] env[61985]: DEBUG oslo_concurrency.lockutils [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.427381] env[61985]: DEBUG oslo_concurrency.lockutils [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.427528] env[61985]: DEBUG nova.compute.manager [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] No waiting events found dispatching network-vif-plugged-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1146.427870] env[61985]: WARNING nova.compute.manager [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received unexpected event network-vif-plugged-eedcafdd-f3f2-4705-a528-258bb48b4574 for instance with vm_state building and task_state spawning. [ 1146.427943] env[61985]: DEBUG nova.compute.manager [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received event network-changed-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1146.428093] env[61985]: DEBUG nova.compute.manager [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Refreshing instance network info cache due to event network-changed-eedcafdd-f3f2-4705-a528-258bb48b4574. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1146.428278] env[61985]: DEBUG oslo_concurrency.lockutils [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] Acquiring lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.465619] env[61985]: DEBUG nova.network.neutron [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updating instance_info_cache with network_info: [{"id": "eedcafdd-f3f2-4705-a528-258bb48b4574", "address": "fa:16:3e:49:21:b1", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeedcafdd-f3", "ovs_interfaceid": "eedcafdd-f3f2-4705-a528-258bb48b4574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.919906] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1146.923259] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936657, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.968624] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.969061] env[61985]: DEBUG nova.compute.manager [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Instance network_info: |[{"id": "eedcafdd-f3f2-4705-a528-258bb48b4574", "address": "fa:16:3e:49:21:b1", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeedcafdd-f3", "ovs_interfaceid": "eedcafdd-f3f2-4705-a528-258bb48b4574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1146.969471] env[61985]: DEBUG oslo_concurrency.lockutils [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] Acquired lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.969718] env[61985]: DEBUG nova.network.neutron [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Refreshing network info cache for port eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1146.971237] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:21:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eedcafdd-f3f2-4705-a528-258bb48b4574', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1146.979815] env[61985]: DEBUG oslo.service.loopingcall [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1146.983620] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1146.984759] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29ceb4a1-d947-4616-af8c-3d98f13577f3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.008936] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1147.008936] env[61985]: value = "task-936658" [ 1147.008936] env[61985]: _type = "Task" [ 1147.008936] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.023402] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936658, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.158740] env[61985]: DEBUG nova.compute.manager [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1147.188690] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1147.188965] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1147.189146] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1147.189343] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1147.189500] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1147.189652] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1147.189872] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1147.190056] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1147.190238] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1147.190410] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1147.190592] env[61985]: DEBUG nova.virt.hardware [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.193750] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c739e7-9c8c-4653-a06a-2457a5b602a2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.202727] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0d0e99-501c-4a3d-8430-77503b6180ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.227568] env[61985]: DEBUG nova.network.neutron [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updated VIF entry in instance network info cache for port eedcafdd-f3f2-4705-a528-258bb48b4574. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1147.227982] env[61985]: DEBUG nova.network.neutron [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updating instance_info_cache with network_info: [{"id": "eedcafdd-f3f2-4705-a528-258bb48b4574", "address": "fa:16:3e:49:21:b1", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeedcafdd-f3", "ovs_interfaceid": "eedcafdd-f3f2-4705-a528-258bb48b4574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.230049] env[61985]: DEBUG nova.network.neutron [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance_info_cache with network_info: [{"id": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "address": "fa:16:3e:d8:5d:8b", "network": {"id": "ddb32e57-13b5-4cb5-8fcb-6a5f28d4573e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-74713238-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5625013599b44418bd56eb604e14be58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7908211b-df93-467b-87a8-3c3d29b03de6", "external-id": "nsx-vlan-transportzone-632", "segmentation_id": 632, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7303cfe7-ec", "ovs_interfaceid": "7303cfe7-ecb8-4ee1-9b69-414635f73c1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.416012] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936657, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.629478} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.416299] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be/8b9791c2-bf5f-495a-8c5e-4e45ff4b91be.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1147.416524] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1147.416778] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eb7d79f8-dd63-4361-9cd3-2877a0af9835 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.425428] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1147.425634] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.296s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.426090] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1147.426090] env[61985]: value = "task-936659" [ 1147.426090] env[61985]: _type = "Task" [ 1147.426090] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.426464] env[61985]: DEBUG oslo_concurrency.lockutils [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.764s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.426642] env[61985]: DEBUG oslo_concurrency.lockutils [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.428780] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.867s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.429022] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.430802] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.430995] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Cleaning up deleted instances {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11232}} [ 1147.442907] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936659, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.464787] env[61985]: INFO nova.scheduler.client.report [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted allocations for instance ad01c86c-4752-4189-a361-afec6fd56495 [ 1147.470316] env[61985]: INFO nova.scheduler.client.report [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted allocations for instance 036fae78-2af0-49f1-85fa-245c29f35ebc [ 1147.520569] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936658, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.534323] env[61985]: DEBUG nova.network.neutron [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Successfully updated port: bb5ccfcb-7752-47cd-b569-3c2861ea2a73 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1147.733748] env[61985]: DEBUG oslo_concurrency.lockutils [req-e8501be6-6a37-44cf-bf37-70452aab498d req-5964a570-b884-4fa0-ae6b-8d73597e17b5 service nova] Releasing lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.733748] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Releasing lock "refresh_cache-c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.733748] env[61985]: DEBUG nova.objects.instance [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lazy-loading 'migration_context' on Instance uuid c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1147.953044] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] There are 62 instances to clean {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1147.953373] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: f2bd7642-e20c-408b-9a7f-73428b89328a] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1147.963215] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936659, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111039} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.963215] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1147.963215] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ecc455b-a268-4aa5-b9b5-fb361005225e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.991170] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be/8b9791c2-bf5f-495a-8c5e-4e45ff4b91be.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1147.991795] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1244699a-bc13-4046-9832-38e5ec32b617 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "ad01c86c-4752-4189-a361-afec6fd56495" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.617s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.993400] env[61985]: DEBUG oslo_concurrency.lockutils [None req-421bd399-0f02-43d0-8158-18a5ea5a589a tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "036fae78-2af0-49f1-85fa-245c29f35ebc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.773s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.994499] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0608833-8263-4919-a569-3bcb7ee18996 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.022532] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936658, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.023886] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1148.023886] env[61985]: value = "task-936660" [ 1148.023886] env[61985]: _type = "Task" [ 1148.023886] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.032622] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936660, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.038208] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "refresh_cache-41627143-d60c-4b18-8192-ace45be71d0d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.038208] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "refresh_cache-41627143-d60c-4b18-8192-ace45be71d0d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.038208] env[61985]: DEBUG nova.network.neutron [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1148.236818] env[61985]: DEBUG nova.objects.base [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1148.237818] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cf5948-a67d-4282-a4e6-15dd2316db3b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.260252] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bcfaf1d-eadc-41cd-90c4-9da3b1b84fca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.268557] env[61985]: DEBUG oslo_vmware.api [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1148.268557] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52602395-e8c5-5209-344d-7a214acecb51" [ 1148.268557] env[61985]: _type = "Task" [ 1148.268557] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.278646] env[61985]: DEBUG oslo_vmware.api [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52602395-e8c5-5209-344d-7a214acecb51, 'name': SearchDatastore_Task, 'duration_secs': 0.008772} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.278970] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.279224] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.456339] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: ef330828-c218-4b80-8b1a-6813bdb67c4d] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1148.521021] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936658, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.537325] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936660, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.559033] env[61985]: DEBUG oslo_vmware.rw_handles [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b80463-fd82-53fd-564a-3c86c0ea5af8/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1148.559643] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753490f5-8b44-46f9-9386-2d3176269beb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.568122] env[61985]: DEBUG oslo_vmware.rw_handles [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b80463-fd82-53fd-564a-3c86c0ea5af8/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1148.568122] env[61985]: ERROR oslo_vmware.rw_handles [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b80463-fd82-53fd-564a-3c86c0ea5af8/disk-0.vmdk due to incomplete transfer. [ 1148.568122] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-efe1f69e-0d59-4fd9-aafa-2498f04d2d1a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.575926] env[61985]: DEBUG oslo_vmware.rw_handles [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b80463-fd82-53fd-564a-3c86c0ea5af8/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1148.576177] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Uploaded image 2bf9b33d-ad4a-42a8-8df6-48e63defc809 to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1148.578642] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1148.579530] env[61985]: DEBUG nova.network.neutron [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1148.581403] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-984ddbf0-a629-4bba-b93d-636ba92b52c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.586439] env[61985]: DEBUG nova.compute.manager [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Received event network-vif-plugged-bb5ccfcb-7752-47cd-b569-3c2861ea2a73 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1148.586439] env[61985]: DEBUG oslo_concurrency.lockutils [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] Acquiring lock "41627143-d60c-4b18-8192-ace45be71d0d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.586651] env[61985]: DEBUG oslo_concurrency.lockutils [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] Lock "41627143-d60c-4b18-8192-ace45be71d0d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.586779] env[61985]: DEBUG oslo_concurrency.lockutils [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] Lock "41627143-d60c-4b18-8192-ace45be71d0d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.587087] env[61985]: DEBUG nova.compute.manager [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] No waiting events found dispatching network-vif-plugged-bb5ccfcb-7752-47cd-b569-3c2861ea2a73 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1148.587204] env[61985]: WARNING nova.compute.manager [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Received unexpected event network-vif-plugged-bb5ccfcb-7752-47cd-b569-3c2861ea2a73 for instance with vm_state building and task_state spawning. [ 1148.587284] env[61985]: DEBUG nova.compute.manager [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Received event network-changed-bb5ccfcb-7752-47cd-b569-3c2861ea2a73 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1148.587446] env[61985]: DEBUG nova.compute.manager [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Refreshing instance network info cache due to event network-changed-bb5ccfcb-7752-47cd-b569-3c2861ea2a73. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1148.587616] env[61985]: DEBUG oslo_concurrency.lockutils [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] Acquiring lock "refresh_cache-41627143-d60c-4b18-8192-ace45be71d0d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.592167] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1148.592167] env[61985]: value = "task-936661" [ 1148.592167] env[61985]: _type = "Task" [ 1148.592167] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.612758] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936661, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.726857] env[61985]: DEBUG nova.network.neutron [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Updating instance_info_cache with network_info: [{"id": "bb5ccfcb-7752-47cd-b569-3c2861ea2a73", "address": "fa:16:3e:e1:55:9e", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb5ccfcb-77", "ovs_interfaceid": "bb5ccfcb-7752-47cd-b569-3c2861ea2a73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.760819] env[61985]: DEBUG oslo_concurrency.lockutils [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.761223] env[61985]: DEBUG oslo_concurrency.lockutils [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.761505] env[61985]: DEBUG oslo_concurrency.lockutils [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.761757] env[61985]: DEBUG oslo_concurrency.lockutils [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.762025] env[61985]: DEBUG oslo_concurrency.lockutils [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.764837] env[61985]: INFO nova.compute.manager [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Terminating instance [ 1148.766861] env[61985]: DEBUG nova.compute.manager [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1148.767131] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1148.768093] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d24ac7-0b52-43a1-91e3-d9c00efa5c7c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.777246] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1148.777512] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ac17cbb-7cbd-44a6-adac-6b4ffbee58ce {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.786000] env[61985]: DEBUG oslo_vmware.api [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1148.786000] env[61985]: value = "task-936662" [ 1148.786000] env[61985]: _type = "Task" [ 1148.786000] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.794085] env[61985]: DEBUG oslo_vmware.api [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936662, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.926971] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c028a5a-6ce7-42b6-8c25-d36748135ec5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.940618] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d2a4b7-4fca-46d0-958f-11511cdc2636 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.972765] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: f45760fa-2298-422e-84ff-83f99b19627d] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1148.975703] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3695ddb-ef84-48ed-82a7-d97847d33b51 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.984775] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3acb2a3-183b-4370-b313-ca38ad6f191e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.004291] env[61985]: DEBUG nova.compute.provider_tree [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1149.022463] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936658, 'name': CreateVM_Task, 'duration_secs': 1.562537} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.022656] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1149.023686] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.024165] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.024247] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1149.024506] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b73e418-b5fc-4c56-a7ee-71aef7520a65 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.036178] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936660, 'name': ReconfigVM_Task, 'duration_secs': 0.961434} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.036461] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1149.036461] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fb8882-d4e4-632e-f043-1255ead2a338" [ 1149.036461] env[61985]: _type = "Task" [ 1149.036461] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.036724] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be/8b9791c2-bf5f-495a-8c5e-4e45ff4b91be.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1149.037405] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea6b2bb3-095f-4053-9fb0-92c7acb58a3e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.050507] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fb8882-d4e4-632e-f043-1255ead2a338, 'name': SearchDatastore_Task, 'duration_secs': 0.011277} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.051787] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.052064] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1149.052317] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.052476] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.052663] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1149.052985] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1149.052985] env[61985]: value = "task-936663" [ 1149.052985] env[61985]: _type = "Task" [ 1149.052985] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.053225] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22d57e06-aad1-45ce-aba2-70fda79eac7a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.064018] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936663, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.066898] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1149.067104] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1149.067858] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a00edae-dc0a-48c8-b5d0-f73750143559 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.074165] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1149.074165] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521a1278-30a9-4b4d-6e6e-ea992edcfa20" [ 1149.074165] env[61985]: _type = "Task" [ 1149.074165] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.082996] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521a1278-30a9-4b4d-6e6e-ea992edcfa20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.102149] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936661, 'name': Destroy_Task, 'duration_secs': 0.477804} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.102447] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Destroyed the VM [ 1149.102717] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1149.103010] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ad0d783f-0f07-481c-af27-e090b59fe54a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.110288] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1149.110288] env[61985]: value = "task-936664" [ 1149.110288] env[61985]: _type = "Task" [ 1149.110288] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.118402] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936664, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.229872] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "refresh_cache-41627143-d60c-4b18-8192-ace45be71d0d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.230422] env[61985]: DEBUG nova.compute.manager [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Instance network_info: |[{"id": "bb5ccfcb-7752-47cd-b569-3c2861ea2a73", "address": "fa:16:3e:e1:55:9e", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb5ccfcb-77", "ovs_interfaceid": "bb5ccfcb-7752-47cd-b569-3c2861ea2a73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1149.230896] env[61985]: DEBUG oslo_concurrency.lockutils [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] Acquired lock "refresh_cache-41627143-d60c-4b18-8192-ace45be71d0d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.231241] env[61985]: DEBUG nova.network.neutron [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Refreshing network info cache for port bb5ccfcb-7752-47cd-b569-3c2861ea2a73 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1149.235647] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:55:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd829efb7-e98e-4b67-bd03-b0888287dbfd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb5ccfcb-7752-47cd-b569-3c2861ea2a73', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1149.247343] env[61985]: DEBUG oslo.service.loopingcall [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1149.251962] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1149.252686] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f921d2c4-944f-46c4-8f80-ff7796ed2860 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.286044] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1149.286044] env[61985]: value = "task-936665" [ 1149.286044] env[61985]: _type = "Task" [ 1149.286044] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.301426] env[61985]: DEBUG oslo_vmware.api [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936662, 'name': PowerOffVM_Task, 'duration_secs': 0.247235} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.305065] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1149.305065] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1149.305278] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936665, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.305477] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce4037a7-340f-491a-a87d-f1031a3d02e3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.479808] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 93dd0394-d7e1-42d5-a58f-fcd621c044b3] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1149.484377] env[61985]: DEBUG nova.network.neutron [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Updated VIF entry in instance network info cache for port bb5ccfcb-7752-47cd-b569-3c2861ea2a73. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1149.484744] env[61985]: DEBUG nova.network.neutron [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Updating instance_info_cache with network_info: [{"id": "bb5ccfcb-7752-47cd-b569-3c2861ea2a73", "address": "fa:16:3e:e1:55:9e", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb5ccfcb-77", "ovs_interfaceid": "bb5ccfcb-7752-47cd-b569-3c2861ea2a73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.507473] env[61985]: DEBUG nova.scheduler.client.report [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1149.566302] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936663, 'name': Rename_Task, 'duration_secs': 0.219209} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.566598] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1149.566857] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a80a63f9-02f0-46db-ba8f-636039aaa185 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.574974] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1149.574974] env[61985]: value = "task-936667" [ 1149.574974] env[61985]: _type = "Task" [ 1149.574974] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.589290] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936667, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.592715] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521a1278-30a9-4b4d-6e6e-ea992edcfa20, 'name': SearchDatastore_Task, 'duration_secs': 0.011161} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.593494] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2683ffd2-36d6-4986-82a3-19f74b9349af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.600702] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1149.600702] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527073f1-0c54-be97-8eb9-485146cc5ccd" [ 1149.600702] env[61985]: _type = "Task" [ 1149.600702] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.609706] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527073f1-0c54-be97-8eb9-485146cc5ccd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.619813] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936664, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.797869] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936665, 'name': CreateVM_Task, 'duration_secs': 0.386779} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.797940] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1149.798711] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.798896] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.799315] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1149.799523] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7dfc0c9-5d96-4017-9342-7f31cb9a17ba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.804957] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1149.804957] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52670bcf-00d7-363e-018a-34058846d587" [ 1149.804957] env[61985]: _type = "Task" [ 1149.804957] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.813481] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52670bcf-00d7-363e-018a-34058846d587, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.983518] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 926a2b59-8614-4e71-9d94-fd9feefdef46] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1149.987358] env[61985]: DEBUG oslo_concurrency.lockutils [req-8453b9e9-b441-401c-a5d9-fd8e113c49de req-005b775b-f163-415b-9fa4-abe4c29ce739 service nova] Releasing lock "refresh_cache-41627143-d60c-4b18-8192-ace45be71d0d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.089880] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936667, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.113228] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "1b0594f2-6191-441f-b373-00b0a925d14d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.113475] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "1b0594f2-6191-441f-b373-00b0a925d14d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.114662] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527073f1-0c54-be97-8eb9-485146cc5ccd, 'name': SearchDatastore_Task, 'duration_secs': 0.011444} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.118266] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.118525] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 45cb51fd-3876-4875-9a5d-b7760072ef55/45cb51fd-3876-4875-9a5d-b7760072ef55.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1150.119079] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-53959c1e-05c2-4b0d-95e0-7327ac32bfb1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.127048] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936664, 'name': RemoveSnapshot_Task, 'duration_secs': 0.75263} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.127939] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1150.128249] env[61985]: DEBUG nova.compute.manager [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1150.130179] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355c3245-8386-4c0d-86f5-4ad80796c8d4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.133205] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1150.133205] env[61985]: value = "task-936668" [ 1150.133205] env[61985]: _type = "Task" [ 1150.133205] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.146256] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936668, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.316387] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52670bcf-00d7-363e-018a-34058846d587, 'name': SearchDatastore_Task, 'duration_secs': 0.010978} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.316708] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.317045] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1150.317220] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.317373] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.317560] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1150.317863] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7479591a-6f82-4906-bdf9-a33d8346f498 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.329685] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1150.329893] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1150.330632] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91fc54b9-7efb-4243-905e-8b840b8cbc63 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.339214] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1150.339214] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5248cc8e-5431-9111-edee-3eab76faf2b6" [ 1150.339214] env[61985]: _type = "Task" [ 1150.339214] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.348405] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5248cc8e-5431-9111-edee-3eab76faf2b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.486611] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 8cfc2489-b3b9-463d-9ad5-0cd73920ed58] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1150.519771] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.240s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.575120] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1150.575371] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1150.575593] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleting the datastore file [datastore2] 59e742d9-c51a-47dd-9ef2-2f070eae94c4 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1150.576263] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b1168d5-b594-4487-909a-13fa65fefc11 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.588843] env[61985]: DEBUG oslo_vmware.api [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936667, 'name': PowerOnVM_Task, 'duration_secs': 0.650498} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.589734] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1150.589983] env[61985]: INFO nova.compute.manager [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Took 9.90 seconds to spawn the instance on the hypervisor. [ 1150.590225] env[61985]: DEBUG nova.compute.manager [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1150.591124] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6a47c9-bc26-4b61-8956-5c956725393f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.595226] env[61985]: DEBUG oslo_vmware.api [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for the task: (returnval){ [ 1150.595226] env[61985]: value = "task-936669" [ 1150.595226] env[61985]: _type = "Task" [ 1150.595226] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.610601] env[61985]: DEBUG oslo_vmware.api [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936669, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.616191] env[61985]: DEBUG nova.compute.manager [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1150.646482] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936668, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5111} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.646946] env[61985]: INFO nova.compute.manager [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Shelve offloading [ 1150.648875] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 45cb51fd-3876-4875-9a5d-b7760072ef55/45cb51fd-3876-4875-9a5d-b7760072ef55.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1150.649116] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1150.649562] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1150.650227] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-31794ee7-8af0-498c-ab3a-0c5736775238 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.652196] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-322f4d6c-32ba-49d5-a2d5-d62b80ab6f06 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.661782] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1150.661782] env[61985]: value = "task-936670" [ 1150.661782] env[61985]: _type = "Task" [ 1150.661782] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.663251] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1150.663251] env[61985]: value = "task-936671" [ 1150.663251] env[61985]: _type = "Task" [ 1150.663251] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.677467] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936670, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.680547] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1150.680739] env[61985]: DEBUG nova.compute.manager [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1150.681506] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181adf66-3600-40f3-ab61-367ffd94a980 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.687693] env[61985]: DEBUG oslo_concurrency.lockutils [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.687846] env[61985]: DEBUG oslo_concurrency.lockutils [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.688070] env[61985]: DEBUG nova.network.neutron [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1150.851143] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5248cc8e-5431-9111-edee-3eab76faf2b6, 'name': SearchDatastore_Task, 'duration_secs': 0.060079} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.852057] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8823429-a5ff-40d1-838b-aa143829817b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.857497] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1150.857497] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52299f0c-b52f-1a29-c3da-c471928cb4bc" [ 1150.857497] env[61985]: _type = "Task" [ 1150.857497] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.865172] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52299f0c-b52f-1a29-c3da-c471928cb4bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.989664] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: acebe0a7-62d4-48d4-a1f7-5df0fa465753] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1151.078786] env[61985]: INFO nova.scheduler.client.report [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted allocation for migration b0648906-0c75-4305-9d80-b664e288823a [ 1151.109622] env[61985]: DEBUG oslo_vmware.api [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Task: {'id': task-936669, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170999} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.110209] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1151.110209] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1151.110350] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1151.110443] env[61985]: INFO nova.compute.manager [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Took 2.34 seconds to destroy the instance on the hypervisor. [ 1151.110701] env[61985]: DEBUG oslo.service.loopingcall [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1151.112489] env[61985]: DEBUG nova.compute.manager [-] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1151.112593] env[61985]: DEBUG nova.network.neutron [-] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1151.114590] env[61985]: INFO nova.compute.manager [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Took 15.55 seconds to build instance. [ 1151.134091] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.134361] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.136037] env[61985]: INFO nova.compute.claims [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1151.174333] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936670, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077354} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.174617] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1151.175398] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21d6a5d-b369-4a4d-923f-fa3f9f37950b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.205563] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 45cb51fd-3876-4875-9a5d-b7760072ef55/45cb51fd-3876-4875-9a5d-b7760072ef55.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1151.206404] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4195a39-6b22-4574-8402-5c98b7a81ae5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.227492] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1151.227492] env[61985]: value = "task-936672" [ 1151.227492] env[61985]: _type = "Task" [ 1151.227492] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.236156] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936672, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.371860] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52299f0c-b52f-1a29-c3da-c471928cb4bc, 'name': SearchDatastore_Task, 'duration_secs': 0.009427} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.377028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.377028] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 41627143-d60c-4b18-8192-ace45be71d0d/41627143-d60c-4b18-8192-ace45be71d0d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1151.377199] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aa9ee02e-db0f-4572-a80c-b49579bf8cb7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.385443] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1151.385443] env[61985]: value = "task-936673" [ 1151.385443] env[61985]: _type = "Task" [ 1151.385443] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.395627] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936673, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.493023] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 9079e605-e865-4eb1-9857-03769e4b5ba6] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1151.586882] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.178s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.590277] env[61985]: DEBUG nova.network.neutron [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating instance_info_cache with network_info: [{"id": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "address": "fa:16:3e:7c:35:40", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03b6ccf4-e2", "ovs_interfaceid": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.617643] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d7c7b626-8128-4346-ae4e-63fa5b06d405 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.074s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.741468] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936672, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.806018] env[61985]: DEBUG nova.compute.manager [req-5cfc8fd8-810a-456e-acd5-464ebd49a977 req-fbe04767-a244-4258-8cfa-a350df717d89 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Received event network-vif-deleted-876d546b-a049-4c69-817f-aa6e7d7a2fc2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1151.806341] env[61985]: INFO nova.compute.manager [req-5cfc8fd8-810a-456e-acd5-464ebd49a977 req-fbe04767-a244-4258-8cfa-a350df717d89 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Neutron deleted interface 876d546b-a049-4c69-817f-aa6e7d7a2fc2; detaching it from the instance and deleting it from the info cache [ 1151.807238] env[61985]: DEBUG nova.network.neutron [req-5cfc8fd8-810a-456e-acd5-464ebd49a977 req-fbe04767-a244-4258-8cfa-a350df717d89 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.876469] env[61985]: DEBUG nova.network.neutron [-] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.898085] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936673, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504497} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.898285] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 41627143-d60c-4b18-8192-ace45be71d0d/41627143-d60c-4b18-8192-ace45be71d0d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1151.898652] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1151.898827] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c33997e2-88ae-40d7-b671-390364ac006b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.908216] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1151.908216] env[61985]: value = "task-936674" [ 1151.908216] env[61985]: _type = "Task" [ 1151.908216] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.917218] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936674, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.996607] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: decfe088-90d6-49f9-9eae-7aa461219eaf] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1152.093599] env[61985]: DEBUG oslo_concurrency.lockutils [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.238494] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936672, 'name': ReconfigVM_Task, 'duration_secs': 0.701252} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.239475] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 45cb51fd-3876-4875-9a5d-b7760072ef55/45cb51fd-3876-4875-9a5d-b7760072ef55.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1152.239475] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb5ceea8-2f78-4ddf-9256-a13466c675be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.246650] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1152.246650] env[61985]: value = "task-936675" [ 1152.246650] env[61985]: _type = "Task" [ 1152.246650] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.255470] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936675, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.266210] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed5f7727-13e6-496e-84e2-68637aa0620b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.273531] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5560d57-2bf3-475a-949b-b47a6b1b42c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.304016] env[61985]: INFO nova.compute.manager [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Rebuilding instance [ 1152.306901] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e515e658-9655-4e05-b586-076cef2c7231 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.309698] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c27c33a-f179-4b04-ae54-c747b6483521 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.318813] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89609795-3174-4325-8f5a-25194497b881 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.325148] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd442637-7d7a-451a-8a04-2368da8ae316 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.350119] env[61985]: DEBUG nova.compute.provider_tree [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.362373] env[61985]: DEBUG nova.compute.manager [req-5cfc8fd8-810a-456e-acd5-464ebd49a977 req-fbe04767-a244-4258-8cfa-a350df717d89 service nova] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Detach interface failed, port_id=876d546b-a049-4c69-817f-aa6e7d7a2fc2, reason: Instance 59e742d9-c51a-47dd-9ef2-2f070eae94c4 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1152.365670] env[61985]: DEBUG nova.compute.manager [req-1ba2f5c0-0b37-49db-8ab6-fae715f9b810 req-3115e17e-62b0-45a3-bf97-99d1bd8cd428 service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received event network-vif-unplugged-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1152.365670] env[61985]: DEBUG oslo_concurrency.lockutils [req-1ba2f5c0-0b37-49db-8ab6-fae715f9b810 req-3115e17e-62b0-45a3-bf97-99d1bd8cd428 service nova] Acquiring lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.365854] env[61985]: DEBUG oslo_concurrency.lockutils [req-1ba2f5c0-0b37-49db-8ab6-fae715f9b810 req-3115e17e-62b0-45a3-bf97-99d1bd8cd428 service nova] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.366080] env[61985]: DEBUG oslo_concurrency.lockutils [req-1ba2f5c0-0b37-49db-8ab6-fae715f9b810 req-3115e17e-62b0-45a3-bf97-99d1bd8cd428 service nova] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.366264] env[61985]: DEBUG nova.compute.manager [req-1ba2f5c0-0b37-49db-8ab6-fae715f9b810 req-3115e17e-62b0-45a3-bf97-99d1bd8cd428 service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] No waiting events found dispatching network-vif-unplugged-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1152.366500] env[61985]: WARNING nova.compute.manager [req-1ba2f5c0-0b37-49db-8ab6-fae715f9b810 req-3115e17e-62b0-45a3-bf97-99d1bd8cd428 service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received unexpected event network-vif-unplugged-03b6ccf4-e23e-472e-a02a-e72bd1686c90 for instance with vm_state shelved and task_state shelving_offloading. [ 1152.378607] env[61985]: DEBUG nova.compute.manager [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1152.379055] env[61985]: INFO nova.compute.manager [-] [instance: 59e742d9-c51a-47dd-9ef2-2f070eae94c4] Took 1.27 seconds to deallocate network for instance. [ 1152.379810] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a6b127-d64b-47c1-b86b-eaec427cdcd6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.419513] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936674, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07782} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.419790] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1152.420585] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d51ed6-efb2-4a11-ae4a-03038038149e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.448099] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 41627143-d60c-4b18-8192-ace45be71d0d/41627143-d60c-4b18-8192-ace45be71d0d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1152.448859] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66d01de1-381f-4e28-aef9-512eaed1c960 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.470489] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1152.470489] env[61985]: value = "task-936676" [ 1152.470489] env[61985]: _type = "Task" [ 1152.470489] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.482271] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936676, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.500468] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 8eea60eb-93cf-4701-9ba3-0805691ca016] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1152.523155] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1152.524275] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3fb1093-17a4-482a-bb4e-264fb03787d9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.533293] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1152.533543] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a6757ce-1653-45b6-9ddc-184d20c6d46f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.757250] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936675, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.865337] env[61985]: DEBUG nova.scheduler.client.report [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1152.890168] env[61985]: DEBUG oslo_concurrency.lockutils [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.892661] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1152.892974] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f150f09c-142e-42e6-9e33-0d4f5b2cb52f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.901742] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1152.901742] env[61985]: value = "task-936678" [ 1152.901742] env[61985]: _type = "Task" [ 1152.901742] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.914518] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936678, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.928142] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.928418] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.928630] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.928862] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.929061] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.931244] env[61985]: INFO nova.compute.manager [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Terminating instance [ 1152.933074] env[61985]: DEBUG nova.compute.manager [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1152.933275] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1152.934083] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998c544b-79bf-4208-ba07-74d30fa87ec0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.942632] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1152.942868] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb98a694-b668-42d6-b487-3544265ed45b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.949662] env[61985]: DEBUG oslo_vmware.api [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1152.949662] env[61985]: value = "task-936679" [ 1152.949662] env[61985]: _type = "Task" [ 1152.949662] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.958266] env[61985]: DEBUG oslo_vmware.api [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936679, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.980457] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936676, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.004228] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 65caa6d1-6e6e-4ab3-bb80-4c7e3b380507] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1153.255047] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1153.255047] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1153.255047] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleting the datastore file [datastore1] 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.255392] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-084265ce-fb9d-4ce7-97c9-7541c43405f6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.266370] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936675, 'name': Rename_Task, 'duration_secs': 0.528795} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.266880] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1153.267177] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-11a6d859-8b6b-4857-9995-a74ebe903a1e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.271545] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1153.271545] env[61985]: value = "task-936680" [ 1153.271545] env[61985]: _type = "Task" [ 1153.271545] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.276785] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1153.276785] env[61985]: value = "task-936681" [ 1153.276785] env[61985]: _type = "Task" [ 1153.276785] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.294639] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936680, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.297789] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936681, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.370351] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.370941] env[61985]: DEBUG nova.compute.manager [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1153.373667] env[61985]: DEBUG oslo_concurrency.lockutils [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.484s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.373903] env[61985]: DEBUG nova.objects.instance [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lazy-loading 'resources' on Instance uuid 59e742d9-c51a-47dd-9ef2-2f070eae94c4 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.412484] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936678, 'name': PowerOffVM_Task, 'duration_secs': 0.23353} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.412753] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1153.412985] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1153.413803] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4090f748-4bb7-44db-ae45-ff9c836602f4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.422661] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1153.423054] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-859ff76f-f51f-4847-8d12-b3e352d2c442 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.458897] env[61985]: DEBUG oslo_vmware.api [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936679, 'name': PowerOffVM_Task, 'duration_secs': 0.362255} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.459226] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1153.459436] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1153.459707] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d58bd7c6-8f0f-4871-bf85-6898d2b7866a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.480917] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936676, 'name': ReconfigVM_Task, 'duration_secs': 0.574928} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.481155] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 41627143-d60c-4b18-8192-ace45be71d0d/41627143-d60c-4b18-8192-ace45be71d0d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1153.481764] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c9db64a-0195-4be0-bcf2-3590816c0c93 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.488101] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1153.488101] env[61985]: value = "task-936684" [ 1153.488101] env[61985]: _type = "Task" [ 1153.488101] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.496911] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936684, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.507545] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 5aedbb9e-8bd1-499a-8195-b3045f7a141f] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1153.516168] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1153.516419] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1153.516610] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleting the datastore file [datastore2] 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.516910] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74f75600-644c-40f1-80d9-b5bc0e92d57a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.526452] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1153.526452] env[61985]: value = "task-936685" [ 1153.526452] env[61985]: _type = "Task" [ 1153.526452] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.538750] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936685, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.544259] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1153.544478] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1153.544659] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleting the datastore file [datastore1] c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.544928] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5957cd91-078f-49f6-b91b-6f69fefc5b3a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.551978] env[61985]: DEBUG oslo_vmware.api [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for the task: (returnval){ [ 1153.551978] env[61985]: value = "task-936686" [ 1153.551978] env[61985]: _type = "Task" [ 1153.551978] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.562343] env[61985]: DEBUG oslo_vmware.api [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936686, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.784169] env[61985]: DEBUG oslo_vmware.api [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936680, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.276945} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.784811] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1153.785027] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1153.785221] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1153.790716] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936681, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.810077] env[61985]: INFO nova.scheduler.client.report [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleted allocations for instance 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 [ 1153.877249] env[61985]: DEBUG nova.compute.utils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1153.882410] env[61985]: DEBUG nova.compute.manager [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1153.882410] env[61985]: DEBUG nova.network.neutron [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1153.930920] env[61985]: DEBUG nova.policy [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b5abdace901430cb1e431c9a933161e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bdb5f76b9b24a7ba4800a88482847da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1153.988665] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f053f56-8200-45bb-bfeb-3d96ab395966 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.002595] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b6e3e9-5f81-4520-a3af-707fdef0f892 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.005759] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936684, 'name': Rename_Task, 'duration_secs': 0.263069} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.006109] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1154.006664] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a374caa-6bae-43e3-beb6-764ad6a9c92c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.035152] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 8c4a1fe4-4ded-4064-8261-24826181a68c] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1154.042628] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991e4645-92c0-4d39-acd6-e4471f0a6b9d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.045224] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1154.045224] env[61985]: value = "task-936687" [ 1154.045224] env[61985]: _type = "Task" [ 1154.045224] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.054635] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936685, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196594} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.056485] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db20523-d653-4b0e-a386-b01acc483d45 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.063716] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1154.063949] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1154.064200] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1154.072382] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936687, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.079214] env[61985]: DEBUG oslo_vmware.api [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Task: {'id': task-936686, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.293726} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.086976] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1154.087601] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1154.087601] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1154.087601] env[61985]: INFO nova.compute.manager [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1154.087855] env[61985]: DEBUG oslo.service.loopingcall [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1154.088314] env[61985]: DEBUG nova.compute.provider_tree [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1154.089518] env[61985]: DEBUG nova.compute.manager [-] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1154.089623] env[61985]: DEBUG nova.network.neutron [-] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1154.289737] env[61985]: DEBUG oslo_vmware.api [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936681, 'name': PowerOnVM_Task, 'duration_secs': 0.978888} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.292298] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1154.292670] env[61985]: INFO nova.compute.manager [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Took 9.56 seconds to spawn the instance on the hypervisor. [ 1154.292903] env[61985]: DEBUG nova.compute.manager [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1154.293834] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e6d315-602b-4c07-8126-6a5b73420517 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.319206] env[61985]: DEBUG nova.network.neutron [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Successfully created port: fdec9b82-21e5-4452-a82d-0c6563b85912 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1154.322222] env[61985]: DEBUG oslo_concurrency.lockutils [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.382134] env[61985]: DEBUG nova.compute.manager [req-91246faf-9022-40ae-816e-ee7df78e9af0 req-21a512b7-cf3b-4e28-af61-e671e633fce7 service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Received event network-vif-deleted-7303cfe7-ecb8-4ee1-9b69-414635f73c1b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1154.382490] env[61985]: INFO nova.compute.manager [req-91246faf-9022-40ae-816e-ee7df78e9af0 req-21a512b7-cf3b-4e28-af61-e671e633fce7 service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Neutron deleted interface 7303cfe7-ecb8-4ee1-9b69-414635f73c1b; detaching it from the instance and deleting it from the info cache [ 1154.382691] env[61985]: DEBUG nova.network.neutron [req-91246faf-9022-40ae-816e-ee7df78e9af0 req-21a512b7-cf3b-4e28-af61-e671e633fce7 service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.384586] env[61985]: DEBUG nova.compute.manager [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1154.403878] env[61985]: DEBUG nova.compute.manager [req-478a409f-c984-4fda-8172-86ec3c15a69a req-0a75a015-9f7a-47ea-b58b-dbe297757987 service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received event network-changed-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1154.404046] env[61985]: DEBUG nova.compute.manager [req-478a409f-c984-4fda-8172-86ec3c15a69a req-0a75a015-9f7a-47ea-b58b-dbe297757987 service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Refreshing instance network info cache due to event network-changed-03b6ccf4-e23e-472e-a02a-e72bd1686c90. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1154.404278] env[61985]: DEBUG oslo_concurrency.lockutils [req-478a409f-c984-4fda-8172-86ec3c15a69a req-0a75a015-9f7a-47ea-b58b-dbe297757987 service nova] Acquiring lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1154.405670] env[61985]: DEBUG oslo_concurrency.lockutils [req-478a409f-c984-4fda-8172-86ec3c15a69a req-0a75a015-9f7a-47ea-b58b-dbe297757987 service nova] Acquired lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.405870] env[61985]: DEBUG nova.network.neutron [req-478a409f-c984-4fda-8172-86ec3c15a69a req-0a75a015-9f7a-47ea-b58b-dbe297757987 service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Refreshing network info cache for port 03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1154.540071] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 5cd5e044-b0da-4564-8c2c-e894eb29a74c] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1154.556876] env[61985]: DEBUG oslo_vmware.api [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936687, 'name': PowerOnVM_Task, 'duration_secs': 0.521741} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.557010] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1154.557222] env[61985]: INFO nova.compute.manager [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Took 7.40 seconds to spawn the instance on the hypervisor. [ 1154.557398] env[61985]: DEBUG nova.compute.manager [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1154.558247] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61c39a3-6ac2-46e1-a8c7-c3a95f8e74c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.593348] env[61985]: DEBUG nova.scheduler.client.report [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1154.816425] env[61985]: INFO nova.compute.manager [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Took 15.83 seconds to build instance. [ 1154.857027] env[61985]: DEBUG nova.network.neutron [-] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.892321] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9342c48-8422-4ac2-9880-dbed192a2e1d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.903573] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efffaec7-3a4d-4b21-a910-2ac894e66b4c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.939783] env[61985]: DEBUG nova.compute.manager [req-91246faf-9022-40ae-816e-ee7df78e9af0 req-21a512b7-cf3b-4e28-af61-e671e633fce7 service nova] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Detach interface failed, port_id=7303cfe7-ecb8-4ee1-9b69-414635f73c1b, reason: Instance c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1155.043449] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 885a36d9-24c6-407a-9d6b-01cd97e90674] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1155.076975] env[61985]: INFO nova.compute.manager [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Took 14.97 seconds to build instance. [ 1155.099329] env[61985]: DEBUG oslo_concurrency.lockutils [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.725s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.102191] env[61985]: DEBUG oslo_concurrency.lockutils [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.779s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.102191] env[61985]: DEBUG nova.objects.instance [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'resources' on Instance uuid 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1155.113168] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1155.113920] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1155.113920] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1155.113920] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1155.114113] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1155.114144] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1155.114411] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1155.114585] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1155.114763] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1155.114974] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1155.115191] env[61985]: DEBUG nova.virt.hardware [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1155.116313] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2def7c41-7f72-4026-a5e2-ad7947547da8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.129153] env[61985]: INFO nova.scheduler.client.report [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Deleted allocations for instance 59e742d9-c51a-47dd-9ef2-2f070eae94c4 [ 1155.131120] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfdc087-2502-4bc2-b4f1-e3aeda86e3fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.151840] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:ba:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f79fbde-374c-4af0-b37b-98c170b5a562', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1155.159419] env[61985]: DEBUG oslo.service.loopingcall [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1155.159419] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1155.159419] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00dc5d6f-4c85-48a4-8dc3-0ffac3e8e27f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.177616] env[61985]: DEBUG nova.network.neutron [req-478a409f-c984-4fda-8172-86ec3c15a69a req-0a75a015-9f7a-47ea-b58b-dbe297757987 service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updated VIF entry in instance network info cache for port 03b6ccf4-e23e-472e-a02a-e72bd1686c90. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1155.177616] env[61985]: DEBUG nova.network.neutron [req-478a409f-c984-4fda-8172-86ec3c15a69a req-0a75a015-9f7a-47ea-b58b-dbe297757987 service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating instance_info_cache with network_info: [{"id": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "address": "fa:16:3e:7c:35:40", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap03b6ccf4-e2", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.183618] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1155.183618] env[61985]: value = "task-936688" [ 1155.183618] env[61985]: _type = "Task" [ 1155.183618] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.192445] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936688, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.318416] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b5d79c81-7f92-429d-b55d-4cc701a5647b tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.341s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.361601] env[61985]: INFO nova.compute.manager [-] [instance: c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f] Took 1.27 seconds to deallocate network for instance. [ 1155.399127] env[61985]: DEBUG nova.compute.manager [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1155.430729] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1155.431332] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1155.431658] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1155.432244] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1155.432553] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1155.433077] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1155.433541] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1155.435063] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1155.435063] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1155.435063] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1155.435063] env[61985]: DEBUG nova.virt.hardware [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1155.436315] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1761a4-3cae-492d-a241-9b558737ee00 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.447350] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce3fb1a-ac01-42c7-9666-b4416ff58eba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.547495] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 8db82979-e527-4c63-8ce8-fe1456f7708c] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1155.582281] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8d3ad593-cb83-4151-a83f-af8355f00b5c tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.486s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.605856] env[61985]: DEBUG nova.objects.instance [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'numa_topology' on Instance uuid 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1155.644533] env[61985]: DEBUG oslo_concurrency.lockutils [None req-028e8067-01e8-4888-bab0-f4a378624224 tempest-ServersTestJSON-532212911 tempest-ServersTestJSON-532212911-project-member] Lock "59e742d9-c51a-47dd-9ef2-2f070eae94c4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.883s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.679883] env[61985]: DEBUG oslo_concurrency.lockutils [req-478a409f-c984-4fda-8172-86ec3c15a69a req-0a75a015-9f7a-47ea-b58b-dbe297757987 service nova] Releasing lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.694754] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936688, 'name': CreateVM_Task, 'duration_secs': 0.423838} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.695011] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1155.695895] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.696122] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.696494] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1155.696796] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d1e14d7-d1fb-4d7f-87a0-32417503d913 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.701886] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1155.701886] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e6e55b-0af8-960e-c615-1709b0653db9" [ 1155.701886] env[61985]: _type = "Task" [ 1155.701886] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.710272] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e6e55b-0af8-960e-c615-1709b0653db9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.868278] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.051408] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 90ac7d3e-aef3-4a50-963a-606e7c58f446] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1156.073561] env[61985]: DEBUG nova.network.neutron [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Successfully updated port: fdec9b82-21e5-4452-a82d-0c6563b85912 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1156.108022] env[61985]: DEBUG nova.objects.base [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Object Instance<0fac6d62-46b4-49a0-ab9a-f654ef27d1f1> lazy-loaded attributes: resources,numa_topology {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1156.215590] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e6e55b-0af8-960e-c615-1709b0653db9, 'name': SearchDatastore_Task, 'duration_secs': 0.01393} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.218558] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.218812] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1156.219061] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.219221] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.219484] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1156.219933] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55682df1-3b3b-4ac7-b889-95d047f592b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.229758] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1156.229758] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1156.230325] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c9c9e01-6adf-4cc3-9d16-ad08d65182b3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.234249] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177c51d9-8c08-4e4f-b1e9-b1d3cdf61966 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.244219] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1156.244219] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e30a96-0311-d912-414c-366b55276f60" [ 1156.244219] env[61985]: _type = "Task" [ 1156.244219] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.245179] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d106be61-c675-4623-8472-ef4d12298c94 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.257934] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e30a96-0311-d912-414c-366b55276f60, 'name': SearchDatastore_Task, 'duration_secs': 0.013703} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.284082] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09574c6a-d45a-4614-a692-31ff5f075c51 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.287037] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4002b5-a92e-47a3-afa6-5814ea2158a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.295222] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1156.295222] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527ef4b0-93dc-bb78-1e1b-487d483e77fc" [ 1156.295222] env[61985]: _type = "Task" [ 1156.295222] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.300191] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac97c63-11f6-4c8f-a35c-0d0a7cf66633 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.314905] env[61985]: DEBUG nova.compute.provider_tree [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1156.320226] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527ef4b0-93dc-bb78-1e1b-487d483e77fc, 'name': SearchDatastore_Task, 'duration_secs': 0.011885} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.320763] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.321044] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be/8b9791c2-bf5f-495a-8c5e-4e45ff4b91be.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1156.321315] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d078dd59-a434-48cc-84b3-c4402decd350 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.329699] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1156.329699] env[61985]: value = "task-936689" [ 1156.329699] env[61985]: _type = "Task" [ 1156.329699] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.338927] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936689, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.410398] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.432433] env[61985]: DEBUG nova.compute.manager [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Received event network-changed-bb5ccfcb-7752-47cd-b569-3c2861ea2a73 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1156.432649] env[61985]: DEBUG nova.compute.manager [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Refreshing instance network info cache due to event network-changed-bb5ccfcb-7752-47cd-b569-3c2861ea2a73. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1156.432882] env[61985]: DEBUG oslo_concurrency.lockutils [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] Acquiring lock "refresh_cache-41627143-d60c-4b18-8192-ace45be71d0d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.433111] env[61985]: DEBUG oslo_concurrency.lockutils [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] Acquired lock "refresh_cache-41627143-d60c-4b18-8192-ace45be71d0d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.433297] env[61985]: DEBUG nova.network.neutron [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Refreshing network info cache for port bb5ccfcb-7752-47cd-b569-3c2861ea2a73 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1156.506696] env[61985]: DEBUG nova.compute.manager [req-3fd31e23-4f77-4242-8040-bdd6b3848b4a req-79761fa5-c25a-4d24-8181-8c87f7301a66 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received event network-changed-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1156.506955] env[61985]: DEBUG nova.compute.manager [req-3fd31e23-4f77-4242-8040-bdd6b3848b4a req-79761fa5-c25a-4d24-8181-8c87f7301a66 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Refreshing instance network info cache due to event network-changed-eedcafdd-f3f2-4705-a528-258bb48b4574. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1156.507158] env[61985]: DEBUG oslo_concurrency.lockutils [req-3fd31e23-4f77-4242-8040-bdd6b3848b4a req-79761fa5-c25a-4d24-8181-8c87f7301a66 service nova] Acquiring lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.507314] env[61985]: DEBUG oslo_concurrency.lockutils [req-3fd31e23-4f77-4242-8040-bdd6b3848b4a req-79761fa5-c25a-4d24-8181-8c87f7301a66 service nova] Acquired lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.507502] env[61985]: DEBUG nova.network.neutron [req-3fd31e23-4f77-4242-8040-bdd6b3848b4a req-79761fa5-c25a-4d24-8181-8c87f7301a66 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Refreshing network info cache for port eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1156.557223] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: bed26b5f-7ca5-405a-884e-02b2495dd977] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1156.576732] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "refresh_cache-1b0594f2-6191-441f-b373-00b0a925d14d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.576944] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "refresh_cache-1b0594f2-6191-441f-b373-00b0a925d14d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.577031] env[61985]: DEBUG nova.network.neutron [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1156.822130] env[61985]: DEBUG nova.scheduler.client.report [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1156.840520] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936689, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.060842] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: b6eb50ff-e685-4e8a-92f5-c2661136c361] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1157.142509] env[61985]: DEBUG nova.network.neutron [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1157.328016] env[61985]: DEBUG oslo_concurrency.lockutils [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.226s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.330772] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.463s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.331123] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.351988] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936689, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.570631} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.355497] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be/8b9791c2-bf5f-495a-8c5e-4e45ff4b91be.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1157.355497] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1157.355497] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db1d265b-90ff-4a5e-a693-b3ac4f16cd8e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.361222] env[61985]: INFO nova.scheduler.client.report [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Deleted allocations for instance c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f [ 1157.370198] env[61985]: DEBUG nova.network.neutron [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Updating instance_info_cache with network_info: [{"id": "fdec9b82-21e5-4452-a82d-0c6563b85912", "address": "fa:16:3e:1b:a0:95", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdec9b82-21", "ovs_interfaceid": "fdec9b82-21e5-4452-a82d-0c6563b85912", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.373031] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1157.373031] env[61985]: value = "task-936690" [ 1157.373031] env[61985]: _type = "Task" [ 1157.373031] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.373893] env[61985]: DEBUG nova.network.neutron [req-3fd31e23-4f77-4242-8040-bdd6b3848b4a req-79761fa5-c25a-4d24-8181-8c87f7301a66 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updated VIF entry in instance network info cache for port eedcafdd-f3f2-4705-a528-258bb48b4574. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1157.374226] env[61985]: DEBUG nova.network.neutron [req-3fd31e23-4f77-4242-8040-bdd6b3848b4a req-79761fa5-c25a-4d24-8181-8c87f7301a66 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updating instance_info_cache with network_info: [{"id": "eedcafdd-f3f2-4705-a528-258bb48b4574", "address": "fa:16:3e:49:21:b1", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeedcafdd-f3", "ovs_interfaceid": "eedcafdd-f3f2-4705-a528-258bb48b4574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.386833] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.565462] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 49707589-4969-4e08-882b-2a2c94bc0d85] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1157.595691] env[61985]: DEBUG nova.network.neutron [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Updated VIF entry in instance network info cache for port bb5ccfcb-7752-47cd-b569-3c2861ea2a73. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1157.596070] env[61985]: DEBUG nova.network.neutron [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Updating instance_info_cache with network_info: [{"id": "bb5ccfcb-7752-47cd-b569-3c2861ea2a73", "address": "fa:16:3e:e1:55:9e", "network": {"id": "97fdc085-4289-4425-b4c5-964ac66654b3", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-553181961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4923009773c047519b883745c59e3516", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb5ccfcb-77", "ovs_interfaceid": "bb5ccfcb-7752-47cd-b569-3c2861ea2a73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.844442] env[61985]: DEBUG oslo_concurrency.lockutils [None req-528a1533-f875-4c82-878d-16a9b80a62cf tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.991s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.846961] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.435s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.846961] env[61985]: INFO nova.compute.manager [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Unshelving [ 1157.869930] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a7322f25-efb2-4c53-948f-e8f8b92c4cf5 tempest-DeleteServersTestJSON-1567781853 tempest-DeleteServersTestJSON-1567781853-project-member] Lock "c5c0cd61-6aaf-4a00-b938-efd6a06f8f3f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.941s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.876865] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "refresh_cache-1b0594f2-6191-441f-b373-00b0a925d14d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.876865] env[61985]: DEBUG nova.compute.manager [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Instance network_info: |[{"id": "fdec9b82-21e5-4452-a82d-0c6563b85912", "address": "fa:16:3e:1b:a0:95", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdec9b82-21", "ovs_interfaceid": "fdec9b82-21e5-4452-a82d-0c6563b85912", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1157.877328] env[61985]: DEBUG oslo_concurrency.lockutils [req-3fd31e23-4f77-4242-8040-bdd6b3848b4a req-79761fa5-c25a-4d24-8181-8c87f7301a66 service nova] Releasing lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.877847] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:a0:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fdec9b82-21e5-4452-a82d-0c6563b85912', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1157.885533] env[61985]: DEBUG oslo.service.loopingcall [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1157.889695] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1157.889925] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7821db8d-0899-4001-b8da-6659e5c6752f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.910194] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078271} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.911564] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1157.912093] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1157.912093] env[61985]: value = "task-936691" [ 1157.912093] env[61985]: _type = "Task" [ 1157.912093] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.912776] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd35cb0-e515-4e76-bf52-654482f48ba1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.923090] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936691, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.940724] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be/8b9791c2-bf5f-495a-8c5e-4e45ff4b91be.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1157.941329] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1734c698-29e3-41dd-b8e6-93056462f868 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.962324] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1157.962324] env[61985]: value = "task-936692" [ 1157.962324] env[61985]: _type = "Task" [ 1157.962324] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.970923] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936692, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.071351] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 5f2aa808-8d1e-471e-9ef7-0f91590ec546] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1158.101065] env[61985]: DEBUG oslo_concurrency.lockutils [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] Releasing lock "refresh_cache-41627143-d60c-4b18-8192-ace45be71d0d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.101065] env[61985]: DEBUG nova.compute.manager [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Received event network-vif-plugged-fdec9b82-21e5-4452-a82d-0c6563b85912 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1158.101065] env[61985]: DEBUG oslo_concurrency.lockutils [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] Acquiring lock "1b0594f2-6191-441f-b373-00b0a925d14d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.101065] env[61985]: DEBUG oslo_concurrency.lockutils [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] Lock "1b0594f2-6191-441f-b373-00b0a925d14d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.101065] env[61985]: DEBUG oslo_concurrency.lockutils [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] Lock "1b0594f2-6191-441f-b373-00b0a925d14d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.101065] env[61985]: DEBUG nova.compute.manager [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] No waiting events found dispatching network-vif-plugged-fdec9b82-21e5-4452-a82d-0c6563b85912 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1158.101065] env[61985]: WARNING nova.compute.manager [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Received unexpected event network-vif-plugged-fdec9b82-21e5-4452-a82d-0c6563b85912 for instance with vm_state building and task_state spawning. [ 1158.101065] env[61985]: DEBUG nova.compute.manager [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Received event network-changed-fdec9b82-21e5-4452-a82d-0c6563b85912 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1158.101065] env[61985]: DEBUG nova.compute.manager [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Refreshing instance network info cache due to event network-changed-fdec9b82-21e5-4452-a82d-0c6563b85912. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1158.101065] env[61985]: DEBUG oslo_concurrency.lockutils [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] Acquiring lock "refresh_cache-1b0594f2-6191-441f-b373-00b0a925d14d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.101065] env[61985]: DEBUG oslo_concurrency.lockutils [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] Acquired lock "refresh_cache-1b0594f2-6191-441f-b373-00b0a925d14d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.101065] env[61985]: DEBUG nova.network.neutron [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Refreshing network info cache for port fdec9b82-21e5-4452-a82d-0c6563b85912 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1158.427685] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936691, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.474910] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936692, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.572969] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 942e0f09-4d34-4aa2-8a60-4c12f899712c] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1158.818461] env[61985]: DEBUG nova.network.neutron [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Updated VIF entry in instance network info cache for port fdec9b82-21e5-4452-a82d-0c6563b85912. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1158.818892] env[61985]: DEBUG nova.network.neutron [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Updating instance_info_cache with network_info: [{"id": "fdec9b82-21e5-4452-a82d-0c6563b85912", "address": "fa:16:3e:1b:a0:95", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdec9b82-21", "ovs_interfaceid": "fdec9b82-21e5-4452-a82d-0c6563b85912", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.860829] env[61985]: DEBUG nova.compute.utils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1158.925671] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936691, 'name': CreateVM_Task, 'duration_secs': 0.573344} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.925671] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1158.926384] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.926547] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.926952] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1158.927450] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34111a28-2a42-4c7f-b175-7982e1ac95dc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.932628] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1158.932628] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529ce854-4760-1c60-7777-e7a275089265" [ 1158.932628] env[61985]: _type = "Task" [ 1158.932628] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.951807] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529ce854-4760-1c60-7777-e7a275089265, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.972019] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936692, 'name': ReconfigVM_Task, 'duration_secs': 0.561263} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.972380] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be/8b9791c2-bf5f-495a-8c5e-4e45ff4b91be.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1158.973051] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62181e14-6fa2-40b5-8988-95a6a5e8493a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.979965] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1158.979965] env[61985]: value = "task-936694" [ 1158.979965] env[61985]: _type = "Task" [ 1158.979965] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.987615] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936694, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.076292] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 172647b8-1f1d-49cb-a0fd-63078bc1ae85] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1159.321698] env[61985]: DEBUG oslo_concurrency.lockutils [req-27dc7856-83e4-4a88-aea0-d2c857054996 req-7f99976c-9e77-4a26-aaff-d6d95498f240 service nova] Releasing lock "refresh_cache-1b0594f2-6191-441f-b373-00b0a925d14d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.363625] env[61985]: INFO nova.virt.block_device [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Booting with volume 9bb2af3d-7489-4dde-b5d2-da06600326e4 at /dev/sdb [ 1159.402791] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c5c324f-dcc3-4888-9735-3420a9cc5a71 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.413108] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061f6db2-398e-4ed0-989d-697b6a73bbe6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.444343] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d07a356f-3c65-4649-8b7a-6dd330205688 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.450263] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]529ce854-4760-1c60-7777-e7a275089265, 'name': SearchDatastore_Task, 'duration_secs': 0.050433} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.450999] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.451273] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1159.451520] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1159.451723] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.452468] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1159.452468] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54ce9a0d-1d85-40ac-a5dc-cffde6330ef4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.458102] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7238f59-709b-4bcc-9822-1d93ebae245b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.470489] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1159.470667] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1159.471756] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23e18876-4ecd-49d2-ae3e-106ca0b28b97 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.476914] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1159.476914] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52597998-4c0a-085c-9a3c-9447947bfe3a" [ 1159.476914] env[61985]: _type = "Task" [ 1159.476914] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.493471] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52597998-4c0a-085c-9a3c-9447947bfe3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.494498] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4976cc-d135-46f8-a6af-e3bd32fda757 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.499398] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936694, 'name': Rename_Task, 'duration_secs': 0.156309} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.499961] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1159.500220] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-716b91d8-daa3-4f8f-88a7-e231580b74a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.503795] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a7edf7-616b-4db5-a2dc-9ffd97e7a6cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.508460] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1159.508460] env[61985]: value = "task-936695" [ 1159.508460] env[61985]: _type = "Task" [ 1159.508460] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.519937] env[61985]: DEBUG nova.virt.block_device [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating existing volume attachment record: 52672e06-3aaf-47af-9db5-0e7e42012303 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1159.525094] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936695, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.579811] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 4da2d122-aa36-4cad-9521-498b53b9f2f6] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1159.964598] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "9eda8ccc-9bfc-4c80-a292-883322afe960" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.964827] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "9eda8ccc-9bfc-4c80-a292-883322afe960" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.989421] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52597998-4c0a-085c-9a3c-9447947bfe3a, 'name': SearchDatastore_Task, 'duration_secs': 0.023468} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.990443] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a845c8c6-fb65-4a6f-a1a5-1c1a7dee8ecf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.999186] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1159.999186] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524215b5-05e8-ae00-1569-12159d6142b3" [ 1159.999186] env[61985]: _type = "Task" [ 1159.999186] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.009579] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524215b5-05e8-ae00-1569-12159d6142b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.018180] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936695, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.082777] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 57591f6f-ed7d-46f3-a9c1-217016c42b1c] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1160.467753] env[61985]: DEBUG nova.compute.manager [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1160.511859] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524215b5-05e8-ae00-1569-12159d6142b3, 'name': SearchDatastore_Task, 'duration_secs': 0.012581} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.515255] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.515557] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 1b0594f2-6191-441f-b373-00b0a925d14d/1b0594f2-6191-441f-b373-00b0a925d14d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1160.516081] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-404359be-2a3a-4f08-8bda-b042359a344a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.525240] env[61985]: DEBUG oslo_vmware.api [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936695, 'name': PowerOnVM_Task, 'duration_secs': 0.711772} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.526571] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1160.526784] env[61985]: DEBUG nova.compute.manager [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1160.527120] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1160.527120] env[61985]: value = "task-936699" [ 1160.527120] env[61985]: _type = "Task" [ 1160.527120] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.527840] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864ea10e-84c6-4280-b41c-33229b61679d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.538372] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936699, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.586383] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 625a3143-d138-4b52-aeb7-9e365c6f1128] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1160.996041] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.996554] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.998946] env[61985]: INFO nova.compute.claims [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1161.048205] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936699, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.050708] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.091056] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 488d0b95-97a9-4193-af62-cc92caf99625] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1161.542364] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936699, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.621983} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.542679] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 1b0594f2-6191-441f-b373-00b0a925d14d/1b0594f2-6191-441f-b373-00b0a925d14d.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1161.542906] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1161.543185] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f220cfa-c784-4c68-8a56-9d34de287e6f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.551596] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1161.551596] env[61985]: value = "task-936700" [ 1161.551596] env[61985]: _type = "Task" [ 1161.551596] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.560383] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936700, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.594309] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: cd8e2eaf-3c13-452b-b2e6-8107f0219378] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1162.065418] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936700, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080496} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.065830] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1162.066955] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e7e5a3-7631-458d-a997-929935d157d7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.104391] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 1b0594f2-6191-441f-b373-00b0a925d14d/1b0594f2-6191-441f-b373-00b0a925d14d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1162.107535] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 62824eac-9412-466a-abcf-1010f6a829e4] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1162.109440] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c55f7d9c-9c9d-443f-bad1-cc1240df60ad {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.135219] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1162.135219] env[61985]: value = "task-936702" [ 1162.135219] env[61985]: _type = "Task" [ 1162.135219] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.149603] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936702, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.199882] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e29493-dcd4-4850-b2f6-672b992edd1d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.208654] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ad394b-aed8-4e9d-9b68-79acb70d9273 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.241677] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12498ab5-a04c-4a8d-80f6-c9209aad9430 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.250351] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedf7022-e61e-451d-8033-6104b862bf99 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.265954] env[61985]: DEBUG nova.compute.provider_tree [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.630271] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 5c2a7072-5b14-4ec7-8060-4e21eea927e3] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1162.647337] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936702, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.654391] env[61985]: DEBUG oslo_concurrency.lockutils [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.654737] env[61985]: DEBUG oslo_concurrency.lockutils [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.654965] env[61985]: DEBUG oslo_concurrency.lockutils [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.655178] env[61985]: DEBUG oslo_concurrency.lockutils [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.655362] env[61985]: DEBUG oslo_concurrency.lockutils [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.657607] env[61985]: INFO nova.compute.manager [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Terminating instance [ 1162.659731] env[61985]: DEBUG nova.compute.manager [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1162.659928] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1162.661341] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d629f143-6f62-41c2-9217-c26e6621d8cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.669232] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1162.669476] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b78d6f2-2a93-4556-a665-e550f42c1812 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.676629] env[61985]: DEBUG oslo_vmware.api [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1162.676629] env[61985]: value = "task-936703" [ 1162.676629] env[61985]: _type = "Task" [ 1162.676629] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.687287] env[61985]: DEBUG oslo_vmware.api [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.769390] env[61985]: DEBUG nova.scheduler.client.report [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1163.132982] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 87006fbf-ea90-4d9a-88af-001de424ac14] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1163.148033] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936702, 'name': ReconfigVM_Task, 'duration_secs': 0.729676} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.148276] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 1b0594f2-6191-441f-b373-00b0a925d14d/1b0594f2-6191-441f-b373-00b0a925d14d.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.149049] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80e3e411-b56f-468f-aa0d-81bfaf511ed5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.156961] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1163.156961] env[61985]: value = "task-936704" [ 1163.156961] env[61985]: _type = "Task" [ 1163.156961] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.165878] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936704, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.187377] env[61985]: DEBUG oslo_vmware.api [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936703, 'name': PowerOffVM_Task, 'duration_secs': 0.270769} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.187698] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1163.187893] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1163.188130] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4942c68-bd99-434c-9547-c2b445bd1388 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.274675] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.275226] env[61985]: DEBUG nova.compute.manager [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1163.278547] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.228s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.278687] env[61985]: DEBUG nova.objects.instance [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61985) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1163.282362] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1163.282573] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1163.282755] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleting the datastore file [datastore2] 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.283269] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34073f31-3534-4183-801d-37028c35e153 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.291749] env[61985]: DEBUG oslo_vmware.api [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1163.291749] env[61985]: value = "task-936706" [ 1163.291749] env[61985]: _type = "Task" [ 1163.291749] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.300730] env[61985]: DEBUG oslo_vmware.api [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936706, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.635027] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: a77ca23f-b2c0-4822-8e48-3e47e0dadb27] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1163.670050] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936704, 'name': Rename_Task, 'duration_secs': 0.187509} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.670490] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1163.670876] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1856869-d800-43f4-ac68-cf22b92b88c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.680763] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1163.680763] env[61985]: value = "task-936707" [ 1163.680763] env[61985]: _type = "Task" [ 1163.680763] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.689674] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936707, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.783948] env[61985]: DEBUG nova.compute.utils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1163.789396] env[61985]: DEBUG nova.compute.manager [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Not allocating networking since 'none' was specified. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1163.802146] env[61985]: DEBUG oslo_vmware.api [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936706, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.139040] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 986c576e-be02-48ac-b24c-72edccab25be] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1164.193135] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936707, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.290856] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8fc0e8a3-8313-4289-813e-6bca7cd144ff tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.292237] env[61985]: DEBUG nova.compute.manager [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1164.305766] env[61985]: DEBUG oslo_vmware.api [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936706, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.662252} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.306154] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.306423] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1164.306660] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1164.306883] env[61985]: INFO nova.compute.manager [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Took 1.65 seconds to destroy the instance on the hypervisor. [ 1164.307193] env[61985]: DEBUG oslo.service.loopingcall [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1164.308232] env[61985]: DEBUG nova.compute.manager [-] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1164.308373] env[61985]: DEBUG nova.network.neutron [-] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1164.597066] env[61985]: DEBUG nova.compute.manager [req-50a1015d-732c-4eae-873e-eeaf7797b741 req-f3cf60b4-cb37-414c-b14b-4335f1164e14 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Received event network-vif-deleted-7f79fbde-374c-4af0-b37b-98c170b5a562 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1164.597460] env[61985]: INFO nova.compute.manager [req-50a1015d-732c-4eae-873e-eeaf7797b741 req-f3cf60b4-cb37-414c-b14b-4335f1164e14 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Neutron deleted interface 7f79fbde-374c-4af0-b37b-98c170b5a562; detaching it from the instance and deleting it from the info cache [ 1164.597939] env[61985]: DEBUG nova.network.neutron [req-50a1015d-732c-4eae-873e-eeaf7797b741 req-f3cf60b4-cb37-414c-b14b-4335f1164e14 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.641938] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: aa088bc3-0fa4-4a01-87a9-cd8b11d9b2e3] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1164.692478] env[61985]: DEBUG oslo_vmware.api [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936707, 'name': PowerOnVM_Task, 'duration_secs': 0.70407} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.692764] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1164.693046] env[61985]: INFO nova.compute.manager [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Took 9.30 seconds to spawn the instance on the hypervisor. [ 1164.693263] env[61985]: DEBUG nova.compute.manager [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1164.694043] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ada439d-6c57-4d9f-a3ea-3355ad12dbf4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.063089] env[61985]: DEBUG nova.network.neutron [-] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.102206] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-61bb4efa-b88e-42ea-8cb6-8d492f5fd8cd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.113198] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b69057e-73d5-4af7-b4a0-57f0d1caffba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.146134] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 6dd8218b-c14d-40c9-87df-097fab06c669] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1165.152021] env[61985]: DEBUG nova.compute.manager [req-50a1015d-732c-4eae-873e-eeaf7797b741 req-f3cf60b4-cb37-414c-b14b-4335f1164e14 service nova] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Detach interface failed, port_id=7f79fbde-374c-4af0-b37b-98c170b5a562, reason: Instance 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1165.152021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.152021] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.152021] env[61985]: DEBUG nova.objects.instance [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'pci_requests' on Instance uuid 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.211169] env[61985]: INFO nova.compute.manager [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Took 14.09 seconds to build instance. [ 1165.305316] env[61985]: DEBUG nova.compute.manager [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1165.334095] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1165.334467] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1165.334750] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1165.335098] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1165.335434] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1165.335912] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1165.335997] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1165.336285] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1165.336571] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1165.336846] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1165.337159] env[61985]: DEBUG nova.virt.hardware [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1165.338391] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52467ecf-3aa2-479d-9832-1407be6d977c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.346499] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98163724-0071-44be-be23-a723e63360b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.360835] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1165.366242] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Creating folder: Project (7a38d55dae784c7f838675fa8e75b4ec). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1165.366523] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aaf4a434-f862-4b15-86e1-6331ba1d1765 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.376638] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Created folder: Project (7a38d55dae784c7f838675fa8e75b4ec) in parent group-v211285. [ 1165.376862] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Creating folder: Instances. Parent ref: group-v211499. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1165.377189] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5727a896-8200-4ef1-8606-b4324bd80c71 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.386521] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Created folder: Instances in parent group-v211499. [ 1165.386734] env[61985]: DEBUG oslo.service.loopingcall [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1165.386938] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1165.387159] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb8cb88b-c7ee-4373-881a-e49fe3fc0d31 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.403730] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1165.403730] env[61985]: value = "task-936710" [ 1165.403730] env[61985]: _type = "Task" [ 1165.403730] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.412176] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936710, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.565834] env[61985]: INFO nova.compute.manager [-] [instance: 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be] Took 1.26 seconds to deallocate network for instance. [ 1165.648829] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 54ba5cfd-185a-4c58-aa5e-83cc096a482e] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1165.654048] env[61985]: DEBUG nova.objects.instance [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'numa_topology' on Instance uuid 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.713173] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c1a6daa8-483e-4fcf-9100-80e2e5295d0c tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "1b0594f2-6191-441f-b373-00b0a925d14d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.599s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.914533] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936710, 'name': CreateVM_Task, 'duration_secs': 0.321429} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.914728] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1165.915254] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.915428] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.915767] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1165.916042] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77a3b1ce-676a-49a6-ac14-a0dc195b6051 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.921762] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1165.921762] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5285f19d-593a-d749-e712-f9ea74a448c3" [ 1165.921762] env[61985]: _type = "Task" [ 1165.921762] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.929984] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5285f19d-593a-d749-e712-f9ea74a448c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.073153] env[61985]: DEBUG oslo_concurrency.lockutils [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.152066] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 518635bf-73ee-404b-ae6a-dc4ee23009d2] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1166.156990] env[61985]: INFO nova.compute.claims [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1166.433060] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5285f19d-593a-d749-e712-f9ea74a448c3, 'name': SearchDatastore_Task, 'duration_secs': 0.011199} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.433411] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.433660] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1166.433910] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.434092] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.434297] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1166.434563] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bac00449-dfad-419f-ad2f-117f2ea647d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.443744] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1166.443920] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1166.444630] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67eaed46-7120-4c00-8e3e-7f98ebe09791 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.449798] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1166.449798] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52662d42-384d-9fa2-cc45-d02545fbc4cf" [ 1166.449798] env[61985]: _type = "Task" [ 1166.449798] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.459011] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52662d42-384d-9fa2-cc45-d02545fbc4cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.626382] env[61985]: DEBUG nova.compute.manager [req-9f7cc97a-89c3-4d32-b273-d06fbfdf8d8f req-513d9668-2483-4e88-a8d4-f5ba5e9be97b service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Received event network-changed-fdec9b82-21e5-4452-a82d-0c6563b85912 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1166.626599] env[61985]: DEBUG nova.compute.manager [req-9f7cc97a-89c3-4d32-b273-d06fbfdf8d8f req-513d9668-2483-4e88-a8d4-f5ba5e9be97b service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Refreshing instance network info cache due to event network-changed-fdec9b82-21e5-4452-a82d-0c6563b85912. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1166.626834] env[61985]: DEBUG oslo_concurrency.lockutils [req-9f7cc97a-89c3-4d32-b273-d06fbfdf8d8f req-513d9668-2483-4e88-a8d4-f5ba5e9be97b service nova] Acquiring lock "refresh_cache-1b0594f2-6191-441f-b373-00b0a925d14d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.626975] env[61985]: DEBUG oslo_concurrency.lockutils [req-9f7cc97a-89c3-4d32-b273-d06fbfdf8d8f req-513d9668-2483-4e88-a8d4-f5ba5e9be97b service nova] Acquired lock "refresh_cache-1b0594f2-6191-441f-b373-00b0a925d14d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.627167] env[61985]: DEBUG nova.network.neutron [req-9f7cc97a-89c3-4d32-b273-d06fbfdf8d8f req-513d9668-2483-4e88-a8d4-f5ba5e9be97b service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Refreshing network info cache for port fdec9b82-21e5-4452-a82d-0c6563b85912 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1166.655265] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: e68cc5ff-83b7-4602-be50-ee4b1fa9ac65] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1166.960845] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52662d42-384d-9fa2-cc45-d02545fbc4cf, 'name': SearchDatastore_Task, 'duration_secs': 0.011114} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.961679] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a625a0f-4e7c-4acc-afba-8579a001e55c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.967833] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1166.967833] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a1c0ba-659d-cd82-b68d-8c9600ce7b62" [ 1166.967833] env[61985]: _type = "Task" [ 1166.967833] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.975435] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a1c0ba-659d-cd82-b68d-8c9600ce7b62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.159833] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: c64056d8-0c22-48db-bf4a-a60f468bdad0] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1167.268030] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3273ecea-77ae-4a45-a5c4-45f55a27bf89 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.276279] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bfedfd-846f-46b5-a5e0-4ff1fff95bb9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.308245] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36fc028e-73c5-44fd-8185-7cb15ee86556 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.316304] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ffc78f5-1f61-4575-8001-d1cde5782f7f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.331388] env[61985]: DEBUG nova.compute.provider_tree [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.343246] env[61985]: DEBUG nova.network.neutron [req-9f7cc97a-89c3-4d32-b273-d06fbfdf8d8f req-513d9668-2483-4e88-a8d4-f5ba5e9be97b service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Updated VIF entry in instance network info cache for port fdec9b82-21e5-4452-a82d-0c6563b85912. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1167.343595] env[61985]: DEBUG nova.network.neutron [req-9f7cc97a-89c3-4d32-b273-d06fbfdf8d8f req-513d9668-2483-4e88-a8d4-f5ba5e9be97b service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Updating instance_info_cache with network_info: [{"id": "fdec9b82-21e5-4452-a82d-0c6563b85912", "address": "fa:16:3e:1b:a0:95", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdec9b82-21", "ovs_interfaceid": "fdec9b82-21e5-4452-a82d-0c6563b85912", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.479385] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a1c0ba-659d-cd82-b68d-8c9600ce7b62, 'name': SearchDatastore_Task, 'duration_secs': 0.011042} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.479654] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.479921] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 9eda8ccc-9bfc-4c80-a292-883322afe960/9eda8ccc-9bfc-4c80-a292-883322afe960.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1167.480247] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-072367bc-6044-4090-b69f-cda9102e5883 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.488889] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1167.488889] env[61985]: value = "task-936712" [ 1167.488889] env[61985]: _type = "Task" [ 1167.488889] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.496808] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936712, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.663113] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 5b3a8ba3-48cb-472a-ab45-4e51af68812c] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1167.835281] env[61985]: DEBUG nova.scheduler.client.report [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1167.846434] env[61985]: DEBUG oslo_concurrency.lockutils [req-9f7cc97a-89c3-4d32-b273-d06fbfdf8d8f req-513d9668-2483-4e88-a8d4-f5ba5e9be97b service nova] Releasing lock "refresh_cache-1b0594f2-6191-441f-b373-00b0a925d14d" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.999287] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936712, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469677} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.999519] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 9eda8ccc-9bfc-4c80-a292-883322afe960/9eda8ccc-9bfc-4c80-a292-883322afe960.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1167.999743] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1167.999999] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f741c001-326d-4bf8-934d-3bc6e90a947a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.007050] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1168.007050] env[61985]: value = "task-936713" [ 1168.007050] env[61985]: _type = "Task" [ 1168.007050] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.016202] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936713, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.167597] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: ce757dda-f58a-47b3-b319-e6b01f05b20b] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1168.341663] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.191s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.344457] env[61985]: DEBUG oslo_concurrency.lockutils [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.271s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.344759] env[61985]: DEBUG nova.objects.instance [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'resources' on Instance uuid 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.372763] env[61985]: INFO nova.network.neutron [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating port 03b6ccf4-e23e-472e-a02a-e72bd1686c90 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1168.520235] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936713, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071025} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.520235] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1168.520235] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9108bd8-cef1-4ed0-935d-85083e5e1d95 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.545437] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 9eda8ccc-9bfc-4c80-a292-883322afe960/9eda8ccc-9bfc-4c80-a292-883322afe960.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1168.545852] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ba7fe78-64c5-4a4c-909b-4312302f939e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.570071] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1168.570071] env[61985]: value = "task-936714" [ 1168.570071] env[61985]: _type = "Task" [ 1168.570071] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.579669] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936714, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.671093] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 17bdf8ec-13d2-459c-bc8e-db6a274fc27e] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1168.971446] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21837344-e1b8-4bf9-9bca-35eab8cbebf8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.989050] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6254d07e-9d5f-477a-bfea-12860d7edabf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.026832] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aecf3e7-d069-4b3b-8345-2fe71edee786 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.040788] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603ec34c-3a93-443d-af40-126f39ef66c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.064224] env[61985]: DEBUG nova.compute.provider_tree [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.082483] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936714, 'name': ReconfigVM_Task, 'duration_secs': 0.361126} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.082821] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 9eda8ccc-9bfc-4c80-a292-883322afe960/9eda8ccc-9bfc-4c80-a292-883322afe960.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1169.083641] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9acecc22-d5df-43fd-b676-e5f5147bac32 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.097263] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1169.097263] env[61985]: value = "task-936715" [ 1169.097263] env[61985]: _type = "Task" [ 1169.097263] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.109037] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936715, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.174890] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 1f284789-1e7b-4e9f-9670-34e8e25cd797] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1169.567638] env[61985]: DEBUG nova.scheduler.client.report [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1169.608159] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936715, 'name': Rename_Task, 'duration_secs': 0.293527} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.608489] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1169.608810] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc09e344-9049-49d2-aa0a-815d7e8df0b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.615495] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1169.615495] env[61985]: value = "task-936717" [ 1169.615495] env[61985]: _type = "Task" [ 1169.615495] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.626651] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936717, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.679270] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: f8924f18-7232-4be6-84bf-1ba05b855ffe] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1170.073159] env[61985]: DEBUG oslo_concurrency.lockutils [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.729s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.095385] env[61985]: INFO nova.scheduler.client.report [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleted allocations for instance 8b9791c2-bf5f-495a-8c5e-4e45ff4b91be [ 1170.125896] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936717, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.182431] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: e82d26b1-5502-4fd2-89c5-ffb0c1557c79] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1170.341022] env[61985]: DEBUG nova.compute.manager [req-470c6e51-6c65-4426-a231-29f43bbc9453 req-a15830a0-1b92-4c50-9aa9-bc406149130d service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received event network-vif-plugged-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1170.341182] env[61985]: DEBUG oslo_concurrency.lockutils [req-470c6e51-6c65-4426-a231-29f43bbc9453 req-a15830a0-1b92-4c50-9aa9-bc406149130d service nova] Acquiring lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.341509] env[61985]: DEBUG oslo_concurrency.lockutils [req-470c6e51-6c65-4426-a231-29f43bbc9453 req-a15830a0-1b92-4c50-9aa9-bc406149130d service nova] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.341715] env[61985]: DEBUG oslo_concurrency.lockutils [req-470c6e51-6c65-4426-a231-29f43bbc9453 req-a15830a0-1b92-4c50-9aa9-bc406149130d service nova] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.341890] env[61985]: DEBUG nova.compute.manager [req-470c6e51-6c65-4426-a231-29f43bbc9453 req-a15830a0-1b92-4c50-9aa9-bc406149130d service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] No waiting events found dispatching network-vif-plugged-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1170.342069] env[61985]: WARNING nova.compute.manager [req-470c6e51-6c65-4426-a231-29f43bbc9453 req-a15830a0-1b92-4c50-9aa9-bc406149130d service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received unexpected event network-vif-plugged-03b6ccf4-e23e-472e-a02a-e72bd1686c90 for instance with vm_state shelved_offloaded and task_state spawning. [ 1170.428707] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.429101] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.429410] env[61985]: DEBUG nova.network.neutron [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1170.602585] env[61985]: DEBUG oslo_concurrency.lockutils [None req-80639e66-19df-47b4-817d-c6e75bcc8aa5 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "8b9791c2-bf5f-495a-8c5e-4e45ff4b91be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.948s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.626908] env[61985]: DEBUG oslo_vmware.api [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936717, 'name': PowerOnVM_Task, 'duration_secs': 0.560493} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.627197] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1170.627403] env[61985]: INFO nova.compute.manager [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Took 5.32 seconds to spawn the instance on the hypervisor. [ 1170.627591] env[61985]: DEBUG nova.compute.manager [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1170.628643] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6d9bd6-0dab-4c55-9e26-d94c9b491632 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.685832] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 95e3e6d7-9203-47a0-bad7-050eba09c511] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1171.146557] env[61985]: INFO nova.compute.manager [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Took 10.17 seconds to build instance. [ 1171.189176] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 83e4ecc0-cc96-48e1-8c1d-7dcf84d72475] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1171.296618] env[61985]: DEBUG nova.network.neutron [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating instance_info_cache with network_info: [{"id": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "address": "fa:16:3e:7c:35:40", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03b6ccf4-e2", "ovs_interfaceid": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.652295] env[61985]: DEBUG oslo_concurrency.lockutils [None req-622be028-c0e1-456c-99b8-ee09314116a8 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "9eda8ccc-9bfc-4c80-a292-883322afe960" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.687s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.692104] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 5173cf26-9a1f-4a0e-a0cb-a4f3ac0663bf] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1171.802324] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1171.806308] env[61985]: INFO nova.compute.manager [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Rebuilding instance [ 1171.829021] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='d99e5917e0596972c7094c16a230b9f2',container_format='bare',created_at=2024-09-18T01:33:44Z,direct_url=,disk_format='vmdk',id=2bf9b33d-ad4a-42a8-8df6-48e63defc809,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1988549478-shelved',owner='e7e22f7d7b7b407ebb2d00e174d8a42c',properties=ImageMetaProps,protected=,size=31662592,status='active',tags=,updated_at=2024-09-18T01:34:00Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1171.829292] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1171.829461] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1171.829656] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1171.829811] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1171.830026] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1171.830212] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1171.830383] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1171.830558] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1171.830803] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1171.830903] env[61985]: DEBUG nova.virt.hardware [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1171.832903] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6844d57-0db9-4fdf-b5fe-16f214311f5d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.848708] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d0394e-6119-4e22-95b5-df1122585be4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.865176] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:35:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e55c248-c504-4c7a-bbe9-f42cf417aee7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03b6ccf4-e23e-472e-a02a-e72bd1686c90', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1171.872898] env[61985]: DEBUG oslo.service.loopingcall [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.875586] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1171.877541] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56b667dc-c35d-4ca7-85c6-6493424989c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.892324] env[61985]: DEBUG nova.compute.manager [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1171.893129] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee41f4ee-924f-46dd-8157-6f871f9c0580 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.902551] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1171.902551] env[61985]: value = "task-936719" [ 1171.902551] env[61985]: _type = "Task" [ 1171.902551] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.910494] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936719, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.196159] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: c4e95f51-9acc-46e8-9921-e0a85fb38598] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1172.369883] env[61985]: DEBUG nova.compute.manager [req-16b8dcb3-889e-49e5-b237-7aaa95982d14 req-36238a01-3033-45c0-83b2-97531b9edcde service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received event network-changed-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1172.370097] env[61985]: DEBUG nova.compute.manager [req-16b8dcb3-889e-49e5-b237-7aaa95982d14 req-36238a01-3033-45c0-83b2-97531b9edcde service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Refreshing instance network info cache due to event network-changed-03b6ccf4-e23e-472e-a02a-e72bd1686c90. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1172.370318] env[61985]: DEBUG oslo_concurrency.lockutils [req-16b8dcb3-889e-49e5-b237-7aaa95982d14 req-36238a01-3033-45c0-83b2-97531b9edcde service nova] Acquiring lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1172.370419] env[61985]: DEBUG oslo_concurrency.lockutils [req-16b8dcb3-889e-49e5-b237-7aaa95982d14 req-36238a01-3033-45c0-83b2-97531b9edcde service nova] Acquired lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.370597] env[61985]: DEBUG nova.network.neutron [req-16b8dcb3-889e-49e5-b237-7aaa95982d14 req-36238a01-3033-45c0-83b2-97531b9edcde service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Refreshing network info cache for port 03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1172.405145] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1172.405956] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a577897-6238-40b8-bfa6-e9e1d5aab5d7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.417069] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936719, 'name': CreateVM_Task, 'duration_secs': 0.364732} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.418329] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1172.418767] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1172.418767] env[61985]: value = "task-936720" [ 1172.418767] env[61985]: _type = "Task" [ 1172.418767] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.419359] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1172.419519] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.419951] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1172.420286] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78bcbd31-a0c1-43c9-8797-22fac26683bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.432494] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936720, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.433430] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1172.433430] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c06846-6638-e282-006f-a1898d5ca132" [ 1172.433430] env[61985]: _type = "Task" [ 1172.433430] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.441546] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c06846-6638-e282-006f-a1898d5ca132, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.701056] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 402ef006-2835-4239-a3c4-1c37a7bba9a5] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1172.725931] env[61985]: DEBUG nova.compute.manager [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Stashing vm_state: active {{(pid=61985) _prep_resize /opt/stack/nova/nova/compute/manager.py:5630}} [ 1172.930260] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936720, 'name': PowerOffVM_Task, 'duration_secs': 0.118159} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.930531] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1172.930761] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1172.931547] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae4fb4e-8834-4fbb-9fe5-cad7753cab51 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.942690] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1172.945816] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2deaec14-8195-46bf-a22a-4c1089bc616e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.947339] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.947581] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Processing image 2bf9b33d-ad4a-42a8-8df6-48e63defc809 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1172.947850] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809/2bf9b33d-ad4a-42a8-8df6-48e63defc809.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1172.948013] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809/2bf9b33d-ad4a-42a8-8df6-48e63defc809.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.948207] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1172.948789] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e2813df3-dad0-46a8-9f36-7c94b7d30d82 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.957468] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1172.957696] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1172.958417] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c046a4d0-fa63-4eea-8f9e-b37c97734dba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.964250] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1172.964250] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5262868a-8466-d87f-babf-9c3b817762ba" [ 1172.964250] env[61985]: _type = "Task" [ 1172.964250] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.973187] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]5262868a-8466-d87f-babf-9c3b817762ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.977327] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1172.977491] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1172.977712] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Deleting the datastore file [datastore1] 9eda8ccc-9bfc-4c80-a292-883322afe960 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1172.977955] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc9f2caf-c1ac-4888-aaa6-a2ee97213672 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.984635] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1172.984635] env[61985]: value = "task-936722" [ 1172.984635] env[61985]: _type = "Task" [ 1172.984635] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.992568] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936722, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.106953] env[61985]: DEBUG nova.network.neutron [req-16b8dcb3-889e-49e5-b237-7aaa95982d14 req-36238a01-3033-45c0-83b2-97531b9edcde service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updated VIF entry in instance network info cache for port 03b6ccf4-e23e-472e-a02a-e72bd1686c90. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1173.107354] env[61985]: DEBUG nova.network.neutron [req-16b8dcb3-889e-49e5-b237-7aaa95982d14 req-36238a01-3033-45c0-83b2-97531b9edcde service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating instance_info_cache with network_info: [{"id": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "address": "fa:16:3e:7c:35:40", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03b6ccf4-e2", "ovs_interfaceid": "03b6ccf4-e23e-472e-a02a-e72bd1686c90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.204524] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 4a94a6f9-1c86-4628-aa63-341f2c114e2a] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1173.245614] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.245919] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.475366] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Preparing fetch location {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1173.475668] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Fetch image to [datastore2] OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641/OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641.vmdk {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1173.475965] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Downloading stream optimized image 2bf9b33d-ad4a-42a8-8df6-48e63defc809 to [datastore2] OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641/OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641.vmdk on the data store datastore2 as vApp {{(pid=61985) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1173.476110] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Downloading image file data 2bf9b33d-ad4a-42a8-8df6-48e63defc809 to the ESX as VM named 'OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641' {{(pid=61985) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1173.494480] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936722, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108863} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.496416] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.496616] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1173.496803] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1173.556380] env[61985]: DEBUG oslo_vmware.rw_handles [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1173.556380] env[61985]: value = "resgroup-9" [ 1173.556380] env[61985]: _type = "ResourcePool" [ 1173.556380] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1173.557315] env[61985]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b39eaf7d-82b1-4e59-9b05-46986d28a0b0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.581539] env[61985]: DEBUG oslo_vmware.rw_handles [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lease: (returnval){ [ 1173.581539] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fb3652-f8aa-115a-e9f1-9a8467fd0e13" [ 1173.581539] env[61985]: _type = "HttpNfcLease" [ 1173.581539] env[61985]: } obtained for vApp import into resource pool (val){ [ 1173.581539] env[61985]: value = "resgroup-9" [ 1173.581539] env[61985]: _type = "ResourcePool" [ 1173.581539] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1173.581930] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the lease: (returnval){ [ 1173.581930] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fb3652-f8aa-115a-e9f1-9a8467fd0e13" [ 1173.581930] env[61985]: _type = "HttpNfcLease" [ 1173.581930] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1173.589703] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1173.589703] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fb3652-f8aa-115a-e9f1-9a8467fd0e13" [ 1173.589703] env[61985]: _type = "HttpNfcLease" [ 1173.589703] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1173.610178] env[61985]: DEBUG oslo_concurrency.lockutils [req-16b8dcb3-889e-49e5-b237-7aaa95982d14 req-36238a01-3033-45c0-83b2-97531b9edcde service nova] Releasing lock "refresh_cache-0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.708267] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: e6a57868-ab6e-45d2-9fa0-94efbb79185b] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1173.750880] env[61985]: INFO nova.compute.claims [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1174.092071] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1174.092071] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fb3652-f8aa-115a-e9f1-9a8467fd0e13" [ 1174.092071] env[61985]: _type = "HttpNfcLease" [ 1174.092071] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1174.092398] env[61985]: DEBUG oslo_vmware.rw_handles [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1174.092398] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52fb3652-f8aa-115a-e9f1-9a8467fd0e13" [ 1174.092398] env[61985]: _type = "HttpNfcLease" [ 1174.092398] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1174.093204] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea139c7-316d-4ccd-bda5-7be640313642 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.102561] env[61985]: DEBUG oslo_vmware.rw_handles [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eb3b3a-c406-8d0a-f75a-352ce5a39680/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1174.102763] env[61985]: DEBUG oslo_vmware.rw_handles [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating HTTP connection to write to file with size = 31662592 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eb3b3a-c406-8d0a-f75a-352ce5a39680/disk-0.vmdk. {{(pid=61985) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1174.165515] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2167da23-08ac-430b-85a1-7648e3415aee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.211632] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 8f6cd002-b3c2-4276-b195-15d09a143d31] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1174.257152] env[61985]: INFO nova.compute.resource_tracker [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating resource usage from migration 4143b7bd-7349-40ce-b676-b590deb22b76 [ 1174.366062] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf176df4-7250-4672-bab5-97c7aacf8f1b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.377160] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667c7946-50ca-4d74-b242-899bfe7229a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.416160] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01da5696-b05e-4526-9fe6-b844161aa264 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.424871] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40bcf662-4b90-427b-b711-73e7f4e5fd7e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.442345] env[61985]: DEBUG nova.compute.provider_tree [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.531477] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1174.531792] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1174.531970] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1174.532224] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1174.532416] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1174.532667] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1174.532861] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1174.533328] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1174.533328] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1174.533503] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1174.533697] env[61985]: DEBUG nova.virt.hardware [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1174.534646] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef24aff7-4aad-466e-819e-d88153e68170 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.548193] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377f6de1-e02b-4bc4-9f20-1654cd4378db {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.564183] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Instance VIF info [] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1174.569906] env[61985]: DEBUG oslo.service.loopingcall [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1174.574017] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1174.574215] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a30f9f7a-5ecb-4f04-a0eb-58490b126be0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.595848] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1174.595848] env[61985]: value = "task-936725" [ 1174.595848] env[61985]: _type = "Task" [ 1174.595848] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.607811] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936725, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.715433] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: cfd59b61-cca9-48d5-85e1-1f45d13f1e88] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1174.948878] env[61985]: DEBUG nova.scheduler.client.report [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1175.106759] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936725, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.219018] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 4aca0ecb-4ae6-4400-accd-d71782b2806d] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1175.262845] env[61985]: DEBUG oslo_vmware.rw_handles [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Completed reading data from the image iterator. {{(pid=61985) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1175.263174] env[61985]: DEBUG oslo_vmware.rw_handles [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eb3b3a-c406-8d0a-f75a-352ce5a39680/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1175.264206] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ece422a-7a8a-4145-9b49-b70977c7764e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.272371] env[61985]: DEBUG oslo_vmware.rw_handles [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eb3b3a-c406-8d0a-f75a-352ce5a39680/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1175.272601] env[61985]: DEBUG oslo_vmware.rw_handles [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eb3b3a-c406-8d0a-f75a-352ce5a39680/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1175.272903] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b9f5d1e4-635a-4ab8-9c2b-db90d0308944 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.456149] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.210s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.456395] env[61985]: INFO nova.compute.manager [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Migrating [ 1175.479947] env[61985]: DEBUG oslo_vmware.rw_handles [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eb3b3a-c406-8d0a-f75a-352ce5a39680/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1175.480292] env[61985]: INFO nova.virt.vmwareapi.images [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Downloaded image file data 2bf9b33d-ad4a-42a8-8df6-48e63defc809 [ 1175.481212] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5436e58-eb2e-42be-8433-d98030e2734a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.502063] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80a180fd-db82-4899-805e-c8f7f9d1d9f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.534939] env[61985]: INFO nova.virt.vmwareapi.images [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] The imported VM was unregistered [ 1175.537732] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Caching image {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1175.537998] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating directory with path [datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809 {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1175.538865] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09b00b18-6a02-4cf0-82a4-f6f2a085285c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.552015] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Created directory with path [datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809 {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1175.552322] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641/OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641.vmdk to [datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809/2bf9b33d-ad4a-42a8-8df6-48e63defc809.vmdk. {{(pid=61985) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1175.552599] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-383e4af1-c460-450c-bafc-cd5abd79d7f6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.560036] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1175.560036] env[61985]: value = "task-936728" [ 1175.560036] env[61985]: _type = "Task" [ 1175.560036] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.569363] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936728, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.614413] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936725, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.723186] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 761d5954-a5ca-4459-a1d6-bfc59b284bf4] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1175.972395] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.972583] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.972768] env[61985]: DEBUG nova.network.neutron [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1176.072338] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936728, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.117619] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936725, 'name': CreateVM_Task, 'duration_secs': 1.319934} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.117839] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1176.118302] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.118555] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.118812] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1176.119096] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0b5cf3d-52f4-4d0d-87ba-80a2f3795645 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.126599] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1176.126599] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52dc5264-2265-1d18-d5c1-1553032e4bf6" [ 1176.126599] env[61985]: _type = "Task" [ 1176.126599] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.138088] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52dc5264-2265-1d18-d5c1-1553032e4bf6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.227452] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: b659ff28-5101-4825-84ea-111351c81145] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1176.573864] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936728, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.640655] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52dc5264-2265-1d18-d5c1-1553032e4bf6, 'name': SearchDatastore_Task, 'duration_secs': 0.064862} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.641166] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.641328] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1176.641578] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.641734] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.641923] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1176.642358] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0cf4b80-d0dd-4734-8d15-350c84ec57e9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.662424] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1176.662667] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1176.663557] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a4a9981-64cd-4f1e-9096-3b0dc011e0aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.673910] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1176.673910] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f4ac0c-755a-db4d-f44c-41e6b9d57907" [ 1176.673910] env[61985]: _type = "Task" [ 1176.673910] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.688126] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f4ac0c-755a-db4d-f44c-41e6b9d57907, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.722957] env[61985]: DEBUG nova.network.neutron [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance_info_cache with network_info: [{"id": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "address": "fa:16:3e:13:20:58", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35f2970-0b", "ovs_interfaceid": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.731340] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: f8bc8b71-0317-479d-b2f9-9471757f0774] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1177.074579] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936728, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.188349] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f4ac0c-755a-db4d-f44c-41e6b9d57907, 'name': SearchDatastore_Task, 'duration_secs': 0.070616} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.189378] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45589833-350b-4ec6-bd36-f3fd580dfac3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.198374] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1177.198374] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52447705-a875-29ad-1d84-af52afb63ea2" [ 1177.198374] env[61985]: _type = "Task" [ 1177.198374] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.209367] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52447705-a875-29ad-1d84-af52afb63ea2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.226286] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.234852] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 3694e20c-ce37-4097-9991-8a06f38b2734] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1177.575088] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936728, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.713488] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52447705-a875-29ad-1d84-af52afb63ea2, 'name': SearchDatastore_Task, 'duration_secs': 0.077155} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.714160] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.714327] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 9eda8ccc-9bfc-4c80-a292-883322afe960/9eda8ccc-9bfc-4c80-a292-883322afe960.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1177.714639] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd69c22f-1320-4071-8727-27a754db4728 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.725746] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1177.725746] env[61985]: value = "task-936729" [ 1177.725746] env[61985]: _type = "Task" [ 1177.725746] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.741214] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 794cff68-6e26-4607-96f7-eaeb41182551] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1177.743351] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936729, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.076429] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936728, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.238597] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936729, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.244639] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: fea0cf39-e851-409f-86f5-31cc128a44dc] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1178.574144] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936728, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.61241} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.574481] env[61985]: INFO nova.virt.vmwareapi.ds_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641/OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641.vmdk to [datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809/2bf9b33d-ad4a-42a8-8df6-48e63defc809.vmdk. [ 1178.574655] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Cleaning up location [datastore2] OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1178.574827] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_712a7f81-600f-46b9-a729-57b73d405641 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1178.575118] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b785e8d-d7ce-4ce2-9f03-c75d611ac755 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.583286] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1178.583286] env[61985]: value = "task-936730" [ 1178.583286] env[61985]: _type = "Task" [ 1178.583286] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.592846] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936730, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.738462] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936729, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.743528] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b172d9-7695-46a2-b40e-6a3e7ed48625 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.747562] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 2322ab72-9841-41fb-9d60-2812baabe108] Instance has had 0 of 5 cleanup attempts {{(pid=61985) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11245}} [ 1178.762628] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance '44361f7b-4609-476f-b4a9-58a7851e6e92' progress to 0 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1179.094071] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936730, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149859} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.094071] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1179.094071] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809/2bf9b33d-ad4a-42a8-8df6-48e63defc809.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.094404] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809/2bf9b33d-ad4a-42a8-8df6-48e63defc809.vmdk to [datastore2] 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1/0fac6d62-46b4-49a0-ab9a-f654ef27d1f1.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1179.094802] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-92442796-d538-4d24-8b0d-20b6b361595d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.102487] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1179.102487] env[61985]: value = "task-936731" [ 1179.102487] env[61985]: _type = "Task" [ 1179.102487] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.110760] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936731, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.237233] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936729, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.018143} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.237575] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 9eda8ccc-9bfc-4c80-a292-883322afe960/9eda8ccc-9bfc-4c80-a292-883322afe960.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1179.237912] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1179.238221] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee7de020-5549-46c4-8061-c3f8595b7998 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.245993] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1179.245993] env[61985]: value = "task-936732" [ 1179.245993] env[61985]: _type = "Task" [ 1179.245993] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.253708] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936732, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.268873] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1179.269186] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.269333] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Cleaning up deleted instances with incomplete migration {{(pid=61985) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11270}} [ 1179.270401] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-774c3c73-84dd-44b4-bc58-2f429f31aa5f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.280158] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1179.280158] env[61985]: value = "task-936733" [ 1179.280158] env[61985]: _type = "Task" [ 1179.280158] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.292097] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936733, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.613310] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936731, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.756527] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936732, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126937} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.756815] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1179.757653] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fd68a0-aaac-4a24-9a10-527edd79f830 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.778386] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 9eda8ccc-9bfc-4c80-a292-883322afe960/9eda8ccc-9bfc-4c80-a292-883322afe960.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1179.778966] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-187dcf47-e603-4679-8362-276960498f6b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.793412] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.804515] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936733, 'name': PowerOffVM_Task, 'duration_secs': 0.202409} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.806032] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1179.806234] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance '44361f7b-4609-476f-b4a9-58a7851e6e92' progress to 17 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1179.810061] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1179.810061] env[61985]: value = "task-936735" [ 1179.810061] env[61985]: _type = "Task" [ 1179.810061] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.819981] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936735, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.115100] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936731, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.313330] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1180.313755] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1180.313896] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1180.314151] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1180.314355] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1180.314577] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1180.314829] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1180.315060] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1180.315373] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1180.315559] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1180.315845] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1180.321486] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac51f983-5eb4-4479-9a50-429755dce2a3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.342299] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936735, 'name': ReconfigVM_Task, 'duration_secs': 0.350089} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.343824] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 9eda8ccc-9bfc-4c80-a292-883322afe960/9eda8ccc-9bfc-4c80-a292-883322afe960.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1180.344605] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1180.344605] env[61985]: value = "task-936736" [ 1180.344605] env[61985]: _type = "Task" [ 1180.344605] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.344864] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd54a9b8-14b2-4d3c-b5e9-2de9832b00d7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.357028] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936736, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.358555] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1180.358555] env[61985]: value = "task-936737" [ 1180.358555] env[61985]: _type = "Task" [ 1180.358555] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.367866] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936737, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.615377] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936731, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.856905] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936736, 'name': ReconfigVM_Task, 'duration_secs': 0.201279} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.857285] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance '44361f7b-4609-476f-b4a9-58a7851e6e92' progress to 33 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1180.870192] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936737, 'name': Rename_Task, 'duration_secs': 0.170011} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.870444] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1180.870704] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64081379-030a-4dff-8bab-52479ed7e4fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.879362] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1180.879362] env[61985]: value = "task-936738" [ 1180.879362] env[61985]: _type = "Task" [ 1180.879362] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.892341] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936738, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.116456] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936731, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.366967] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1181.367278] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1181.367431] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1181.367619] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1181.367801] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1181.370561] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1181.370561] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1181.370561] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1181.370561] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1181.370561] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1181.370561] env[61985]: DEBUG nova.virt.hardware [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1181.374098] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Reconfiguring VM instance instance-00000057 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1181.374411] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d03b44ef-db3d-4f94-bd2b-d7192a216800 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.395926] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936738, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.397137] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1181.397137] env[61985]: value = "task-936739" [ 1181.397137] env[61985]: _type = "Task" [ 1181.397137] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.404595] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936739, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.615982] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936731, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.26498} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.616295] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2bf9b33d-ad4a-42a8-8df6-48e63defc809/2bf9b33d-ad4a-42a8-8df6-48e63defc809.vmdk to [datastore2] 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1/0fac6d62-46b4-49a0-ab9a-f654ef27d1f1.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1181.617088] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0570f4-f14b-4f8c-b278-b08ed2146eed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.640171] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1/0fac6d62-46b4-49a0-ab9a-f654ef27d1f1.vmdk or device None with type streamOptimized {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1181.640495] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e81c5d1-6076-4fbf-a261-7377736b6775 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.664262] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1181.664262] env[61985]: value = "task-936741" [ 1181.664262] env[61985]: _type = "Task" [ 1181.664262] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.673338] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936741, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.897779] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936738, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.905716] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936739, 'name': ReconfigVM_Task, 'duration_secs': 0.178468} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.906014] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Reconfigured VM instance instance-00000057 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1181.906811] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c470d3f2-eb94-4f82-a71f-a80dc221a748 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.930733] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 44361f7b-4609-476f-b4a9-58a7851e6e92/44361f7b-4609-476f-b4a9-58a7851e6e92.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1181.931110] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fa26887-e241-4bd0-8edb-35a311f44945 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.952935] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1181.952935] env[61985]: value = "task-936742" [ 1181.952935] env[61985]: _type = "Task" [ 1181.952935] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.963014] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936742, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.174854] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936741, 'name': ReconfigVM_Task, 'duration_secs': 0.319711} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.175132] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1/0fac6d62-46b4-49a0-ab9a-f654ef27d1f1.vmdk or device None with type streamOptimized {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.176400] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_type': 'disk', 'encryption_secret_uuid': None, 'encryption_options': None, 'encryption_format': None, 'disk_bus': None, 'device_name': '/dev/sda', 'boot_index': 0, 'encrypted': False, 'size': 0, 'guest_format': None, 'image_id': 'fe6d817b-6194-440f-988a-f4a94c580922'}], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211498', 'volume_id': '9bb2af3d-7489-4dde-b5d2-da06600326e4', 'name': 'volume-9bb2af3d-7489-4dde-b5d2-da06600326e4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '0fac6d62-46b4-49a0-ab9a-f654ef27d1f1', 'attached_at': '', 'detached_at': '', 'volume_id': '9bb2af3d-7489-4dde-b5d2-da06600326e4', 'serial': '9bb2af3d-7489-4dde-b5d2-da06600326e4'}, 'device_type': None, 'attachment_id': '52672e06-3aaf-47af-9db5-0e7e42012303', 'mount_device': '/dev/sdb', 'disk_bus': None, 'delete_on_termination': False, 'boot_index': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=61985) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1182.176612] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1182.176826] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211498', 'volume_id': '9bb2af3d-7489-4dde-b5d2-da06600326e4', 'name': 'volume-9bb2af3d-7489-4dde-b5d2-da06600326e4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '0fac6d62-46b4-49a0-ab9a-f654ef27d1f1', 'attached_at': '', 'detached_at': '', 'volume_id': '9bb2af3d-7489-4dde-b5d2-da06600326e4', 'serial': '9bb2af3d-7489-4dde-b5d2-da06600326e4'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1182.177686] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fccff68f-b680-4eb3-b1a1-62a07f641b41 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.195232] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d56a00b-0d7b-4cbf-bef2-b1a8c0925205 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.219906] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] volume-9bb2af3d-7489-4dde-b5d2-da06600326e4/volume-9bb2af3d-7489-4dde-b5d2-da06600326e4.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1182.220223] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bd920ab-53f9-4257-b4a0-c3925bc450a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.238247] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1182.238247] env[61985]: value = "task-936743" [ 1182.238247] env[61985]: _type = "Task" [ 1182.238247] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.246276] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936743, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.397243] env[61985]: DEBUG oslo_vmware.api [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936738, 'name': PowerOnVM_Task, 'duration_secs': 1.119429} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.398235] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1182.398370] env[61985]: DEBUG nova.compute.manager [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1182.399183] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5547b0ab-6f28-4231-b206-9657e62489cd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.461994] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936742, 'name': ReconfigVM_Task, 'duration_secs': 0.265599} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.462234] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 44361f7b-4609-476f-b4a9-58a7851e6e92/44361f7b-4609-476f-b4a9-58a7851e6e92.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.462505] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance '44361f7b-4609-476f-b4a9-58a7851e6e92' progress to 50 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1182.749427] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936743, 'name': ReconfigVM_Task, 'duration_secs': 0.331167} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.749730] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Reconfigured VM instance instance-00000066 to attach disk [datastore2] volume-9bb2af3d-7489-4dde-b5d2-da06600326e4/volume-9bb2af3d-7489-4dde-b5d2-da06600326e4.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.754606] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fc24277-0798-4e15-9981-5d4e7201b3fc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.771232] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1182.771232] env[61985]: value = "task-936744" [ 1182.771232] env[61985]: _type = "Task" [ 1182.771232] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.780025] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936744, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.917202] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.917325] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.917477] env[61985]: DEBUG nova.objects.instance [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61985) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1182.969105] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09ec015-6f09-4c58-891b-fb925f1ae53e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.988704] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac083e05-50d7-4bc1-a6ef-38c4165ec2d5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.007117] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance '44361f7b-4609-476f-b4a9-58a7851e6e92' progress to 67 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1183.283709] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936744, 'name': ReconfigVM_Task, 'duration_secs': 0.195333} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.284022] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211498', 'volume_id': '9bb2af3d-7489-4dde-b5d2-da06600326e4', 'name': 'volume-9bb2af3d-7489-4dde-b5d2-da06600326e4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '0fac6d62-46b4-49a0-ab9a-f654ef27d1f1', 'attached_at': '', 'detached_at': '', 'volume_id': '9bb2af3d-7489-4dde-b5d2-da06600326e4', 'serial': '9bb2af3d-7489-4dde-b5d2-da06600326e4'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1183.284611] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6df7dba-f50e-433a-80b2-72c0eea49762 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.291692] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1183.291692] env[61985]: value = "task-936745" [ 1183.291692] env[61985]: _type = "Task" [ 1183.291692] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.300086] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936745, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.546033] env[61985]: DEBUG nova.network.neutron [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Port e35f2970-0b73-4b31-925f-56ae8e0e7f39 binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1183.640843] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "9eda8ccc-9bfc-4c80-a292-883322afe960" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.641237] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "9eda8ccc-9bfc-4c80-a292-883322afe960" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.641506] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "9eda8ccc-9bfc-4c80-a292-883322afe960-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.641736] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "9eda8ccc-9bfc-4c80-a292-883322afe960-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.641968] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "9eda8ccc-9bfc-4c80-a292-883322afe960-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.644228] env[61985]: INFO nova.compute.manager [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Terminating instance [ 1183.646143] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "refresh_cache-9eda8ccc-9bfc-4c80-a292-883322afe960" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.646322] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquired lock "refresh_cache-9eda8ccc-9bfc-4c80-a292-883322afe960" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.646519] env[61985]: DEBUG nova.network.neutron [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1183.802601] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936745, 'name': Rename_Task, 'duration_secs': 0.162374} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.803611] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1183.803718] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-297715ad-8cc5-40d8-8549-956292f010e5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.811219] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1183.811219] env[61985]: value = "task-936747" [ 1183.811219] env[61985]: _type = "Task" [ 1183.811219] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.819308] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936747, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.925714] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ff6cdf10-6178-4cf7-9df3-4d6d533d77fc tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.165526] env[61985]: DEBUG nova.network.neutron [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1184.215065] env[61985]: DEBUG nova.network.neutron [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.322617] env[61985]: DEBUG oslo_vmware.api [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936747, 'name': PowerOnVM_Task, 'duration_secs': 0.490047} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.322929] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1184.421564] env[61985]: DEBUG nova.compute.manager [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.422833] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7912451-253b-4560-bcbf-46696aaa648d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.571321] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "44361f7b-4609-476f-b4a9-58a7851e6e92-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.571831] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.572062] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.717852] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Releasing lock "refresh_cache-9eda8ccc-9bfc-4c80-a292-883322afe960" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.718343] env[61985]: DEBUG nova.compute.manager [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1184.718552] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1184.719510] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e54ba8-0d30-4772-b52e-5d46fca5e778 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.727808] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1184.728139] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5256c73-8318-4afb-90e5-04521f2082f9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.736275] env[61985]: DEBUG oslo_vmware.api [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1184.736275] env[61985]: value = "task-936748" [ 1184.736275] env[61985]: _type = "Task" [ 1184.736275] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.745196] env[61985]: DEBUG oslo_vmware.api [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936748, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.942717] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dc59483b-1801-4e2b-bd91-0e13b12069ab tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 27.097s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.247205] env[61985]: DEBUG oslo_vmware.api [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936748, 'name': PowerOffVM_Task, 'duration_secs': 0.133895} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.247507] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1185.247717] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1185.247975] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc0bb31e-2bf1-4855-97b2-d1efeef8a083 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.277326] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1185.277583] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1185.277803] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Deleting the datastore file [datastore2] 9eda8ccc-9bfc-4c80-a292-883322afe960 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1185.278106] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d277595-7a9e-4c35-a293-caf302b2dee3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.284790] env[61985]: DEBUG oslo_vmware.api [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for the task: (returnval){ [ 1185.284790] env[61985]: value = "task-936750" [ 1185.284790] env[61985]: _type = "Task" [ 1185.284790] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.293713] env[61985]: DEBUG oslo_vmware.api [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936750, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.608986] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.609281] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.609373] env[61985]: DEBUG nova.network.neutron [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1185.795741] env[61985]: DEBUG oslo_vmware.api [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Task: {'id': task-936750, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22201} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.796063] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1185.796304] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1185.796530] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1185.796725] env[61985]: INFO nova.compute.manager [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1185.796993] env[61985]: DEBUG oslo.service.loopingcall [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1185.797195] env[61985]: DEBUG nova.compute.manager [-] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1185.797290] env[61985]: DEBUG nova.network.neutron [-] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1185.813064] env[61985]: DEBUG nova.network.neutron [-] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1186.316174] env[61985]: DEBUG nova.network.neutron [-] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.586764] env[61985]: DEBUG nova.network.neutron [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance_info_cache with network_info: [{"id": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "address": "fa:16:3e:13:20:58", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35f2970-0b", "ovs_interfaceid": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.808057] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "41627143-d60c-4b18-8192-ace45be71d0d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.808057] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.822679] env[61985]: INFO nova.compute.manager [-] [instance: 9eda8ccc-9bfc-4c80-a292-883322afe960] Took 1.03 seconds to deallocate network for instance. [ 1187.090607] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.311746] env[61985]: DEBUG nova.compute.utils [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1187.333374] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.333693] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.333981] env[61985]: DEBUG nova.objects.instance [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lazy-loading 'resources' on Instance uuid 9eda8ccc-9bfc-4c80-a292-883322afe960 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.611764] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f587939c-308c-44d2-ab6f-bc3bc3108ed1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.631414] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82fb649f-6855-40c0-9d9d-8a878d463d2d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.638449] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance '44361f7b-4609-476f-b4a9-58a7851e6e92' progress to 83 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1187.752295] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "f544bc65-af75-4701-802c-f6f39f39e352" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.752606] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "f544bc65-af75-4701-802c-f6f39f39e352" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.814434] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.941602] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b156c759-3907-4ac9-ad7c-ef4499325ac5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.949330] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d9748f-07d9-4be3-a493-01b86856b6f7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.981151] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4208b6ab-3658-44b1-ab27-b21e87dbcf1e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.989528] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9e8c8c-d9e2-4d6f-a515-09249e2823d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.002757] env[61985]: DEBUG nova.compute.provider_tree [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.144602] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1188.144843] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9689211-8970-40f7-99fa-be5a3beac629 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.153767] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1188.153767] env[61985]: value = "task-936751" [ 1188.153767] env[61985]: _type = "Task" [ 1188.153767] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.161476] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936751, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.255065] env[61985]: DEBUG nova.compute.manager [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1188.506283] env[61985]: DEBUG nova.scheduler.client.report [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1188.664173] env[61985]: DEBUG oslo_vmware.api [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936751, 'name': PowerOnVM_Task, 'duration_secs': 0.373297} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.664452] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1188.664647] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c26d9004-d72d-4a7d-99b8-2a67fc3d347c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance '44361f7b-4609-476f-b4a9-58a7851e6e92' progress to 100 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1188.774147] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.874718] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "41627143-d60c-4b18-8192-ace45be71d0d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.875099] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.875291] env[61985]: INFO nova.compute.manager [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Attaching volume db955da0-6604-4fdd-85d1-b5482e8a6d9f to /dev/sdb [ 1188.934328] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7fa07c-8c13-4f5d-a1dd-3b76d2c67b9c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.942785] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48dcf00f-b52c-4637-91bd-b06c34c0c744 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.959069] env[61985]: DEBUG nova.virt.block_device [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Updating existing volume attachment record: 77640a34-6061-4bff-b93c-d82e18339c02 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1189.012491] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.679s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.014762] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.241s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.016302] env[61985]: INFO nova.compute.claims [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1189.034791] env[61985]: INFO nova.scheduler.client.report [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Deleted allocations for instance 9eda8ccc-9bfc-4c80-a292-883322afe960 [ 1189.541824] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4a20e00e-65f1-4a3f-b9ec-964fa0c8de00 tempest-ServerShowV257Test-950507678 tempest-ServerShowV257Test-950507678-project-member] Lock "9eda8ccc-9bfc-4c80-a292-883322afe960" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.900s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.129177] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39378309-d994-47ce-b232-c2684ff7f657 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.137451] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea53bcb-0341-4724-ae12-cefeb701fb36 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.169177] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089d3a0a-d602-4c52-bd99-573a3bb869f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.176999] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7fa6da-6a63-4794-95f5-60404e8068a2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.197529] env[61985]: DEBUG nova.compute.provider_tree [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.427528] env[61985]: DEBUG oslo_concurrency.lockutils [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "44361f7b-4609-476f-b4a9-58a7851e6e92" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.428529] env[61985]: DEBUG oslo_concurrency.lockutils [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.428773] env[61985]: DEBUG nova.compute.manager [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Going to confirm migration 8 {{(pid=61985) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4789}} [ 1190.700295] env[61985]: DEBUG nova.scheduler.client.report [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.959632] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.959914] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1191.086900] env[61985]: DEBUG oslo_concurrency.lockutils [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.087123] env[61985]: DEBUG oslo_concurrency.lockutils [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.087315] env[61985]: DEBUG nova.network.neutron [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1191.087512] env[61985]: DEBUG nova.objects.instance [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'info_cache' on Instance uuid 44361f7b-4609-476f-b4a9-58a7851e6e92 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1191.205488] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.191s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.205972] env[61985]: DEBUG nova.compute.manager [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1191.466128] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1191.466298] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1191.711798] env[61985]: DEBUG nova.compute.utils [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1191.713131] env[61985]: DEBUG nova.compute.manager [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1191.714287] env[61985]: DEBUG nova.network.neutron [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1191.758927] env[61985]: DEBUG nova.policy [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b5abdace901430cb1e431c9a933161e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bdb5f76b9b24a7ba4800a88482847da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1192.038098] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.038449] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquired lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.038627] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 073a7668-39e6-480d-9350-835a0282b456] Forcefully refreshing network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1192.150475] env[61985]: DEBUG nova.network.neutron [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Successfully created port: d3381564-0201-44c0-be83-9d7de95b0fce {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1192.216784] env[61985]: DEBUG nova.compute.manager [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1192.313943] env[61985]: DEBUG nova.network.neutron [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance_info_cache with network_info: [{"id": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "address": "fa:16:3e:13:20:58", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape35f2970-0b", "ovs_interfaceid": "e35f2970-0b73-4b31-925f-56ae8e0e7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.527319] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "45cb51fd-3876-4875-9a5d-b7760072ef55" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.527319] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.527319] env[61985]: INFO nova.compute.manager [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Shelving [ 1192.724254] env[61985]: INFO nova.virt.block_device [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Booting with volume 8253a6a8-8fe6-4148-adf3-55531c7a94c3 at /dev/sda [ 1192.756294] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c60a45e6-bde5-4290-939d-2909c50e9b7c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.766665] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e04731-b4ec-4ffa-a10d-75383dd29ad4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.796560] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f63810b1-622f-4144-984f-dd34bc8c47ed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.805746] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fcf3cd4-5fd0-443c-a3b2-9ffa8348391a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.816150] env[61985]: DEBUG oslo_concurrency.lockutils [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-44361f7b-4609-476f-b4a9-58a7851e6e92" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.816409] env[61985]: DEBUG nova.objects.instance [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'migration_context' on Instance uuid 44361f7b-4609-476f-b4a9-58a7851e6e92 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.836635] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf84a69-8a6b-4785-b020-e514e528ca6e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.843673] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a442d7-319e-4f02-9446-e8dabd1dab6e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.857527] env[61985]: DEBUG nova.virt.block_device [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating existing volume attachment record: 7d6ad0a0-e18d-47d1-9fab-7be7646ffef4 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1193.034082] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1193.034366] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee43ffa5-b1d8-4011-b503-396254948f56 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.042499] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1193.042499] env[61985]: value = "task-936754" [ 1193.042499] env[61985]: _type = "Task" [ 1193.042499] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.052421] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936754, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.271852] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 073a7668-39e6-480d-9350-835a0282b456] Updating instance_info_cache with network_info: [{"id": "a5e0f361-c876-4212-8c4a-dfdfb17d5d4f", "address": "fa:16:3e:3a:9a:10", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5e0f361-c8", "ovs_interfaceid": "a5e0f361-c876-4212-8c4a-dfdfb17d5d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.319787] env[61985]: DEBUG nova.objects.base [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Object Instance<44361f7b-4609-476f-b4a9-58a7851e6e92> lazy-loaded attributes: info_cache,migration_context {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1193.322877] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397563f8-66b7-4ede-9028-ee2f032bb120 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.346075] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8664a41b-5253-49d8-bc1b-6b329ed7588b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.353364] env[61985]: DEBUG oslo_vmware.api [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1193.353364] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e2b8ae-7c10-f0ce-ad62-052c844b74fd" [ 1193.353364] env[61985]: _type = "Task" [ 1193.353364] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.364349] env[61985]: DEBUG oslo_vmware.api [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e2b8ae-7c10-f0ce-ad62-052c844b74fd, 'name': SearchDatastore_Task, 'duration_secs': 0.00806} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.364637] env[61985]: DEBUG oslo_concurrency.lockutils [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.364875] env[61985]: DEBUG oslo_concurrency.lockutils [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.504655] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1193.504902] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211507', 'volume_id': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'name': 'volume-db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41627143-d60c-4b18-8192-ace45be71d0d', 'attached_at': '', 'detached_at': '', 'volume_id': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'serial': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1193.506395] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd98ba5f-750d-4535-ab31-1f1b56aade72 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.523133] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba21c66-fa3a-46dd-a7ad-dfa01e8e0ace {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.551230] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] volume-db955da0-6604-4fdd-85d1-b5482e8a6d9f/volume-db955da0-6604-4fdd-85d1-b5482e8a6d9f.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1193.551230] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26872132-b076-48c7-ae45-03b113bcdb9d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.572525] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936754, 'name': PowerOffVM_Task, 'duration_secs': 0.370313} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.573882] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1193.574267] env[61985]: DEBUG oslo_vmware.api [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1193.574267] env[61985]: value = "task-936755" [ 1193.574267] env[61985]: _type = "Task" [ 1193.574267] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.577506] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a3b605-715a-4c1d-a37e-d459d2cd7c91 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.587084] env[61985]: DEBUG oslo_vmware.api [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936755, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.604029] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a118a327-9802-4a9c-9be8-f8f1ff952025 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.625961] env[61985]: DEBUG nova.compute.manager [req-71471e06-e98d-4bc3-afb9-830d3ac6e97f req-cc891684-a05d-453f-b117-1ac5f5e711bf service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Received event network-vif-plugged-d3381564-0201-44c0-be83-9d7de95b0fce {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1193.626211] env[61985]: DEBUG oslo_concurrency.lockutils [req-71471e06-e98d-4bc3-afb9-830d3ac6e97f req-cc891684-a05d-453f-b117-1ac5f5e711bf service nova] Acquiring lock "f544bc65-af75-4701-802c-f6f39f39e352-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.626439] env[61985]: DEBUG oslo_concurrency.lockutils [req-71471e06-e98d-4bc3-afb9-830d3ac6e97f req-cc891684-a05d-453f-b117-1ac5f5e711bf service nova] Lock "f544bc65-af75-4701-802c-f6f39f39e352-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.626786] env[61985]: DEBUG oslo_concurrency.lockutils [req-71471e06-e98d-4bc3-afb9-830d3ac6e97f req-cc891684-a05d-453f-b117-1ac5f5e711bf service nova] Lock "f544bc65-af75-4701-802c-f6f39f39e352-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.627011] env[61985]: DEBUG nova.compute.manager [req-71471e06-e98d-4bc3-afb9-830d3ac6e97f req-cc891684-a05d-453f-b117-1ac5f5e711bf service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] No waiting events found dispatching network-vif-plugged-d3381564-0201-44c0-be83-9d7de95b0fce {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1193.627254] env[61985]: WARNING nova.compute.manager [req-71471e06-e98d-4bc3-afb9-830d3ac6e97f req-cc891684-a05d-453f-b117-1ac5f5e711bf service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Received unexpected event network-vif-plugged-d3381564-0201-44c0-be83-9d7de95b0fce for instance with vm_state building and task_state block_device_mapping. [ 1193.754330] env[61985]: DEBUG nova.network.neutron [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Successfully updated port: d3381564-0201-44c0-be83-9d7de95b0fce {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1193.775268] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Releasing lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.775481] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 073a7668-39e6-480d-9350-835a0282b456] Updated the network info_cache for instance {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10003}} [ 1193.775674] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.775838] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.776017] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.776157] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.776303] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.776440] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.981686] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc0fa0b-8915-4291-8817-40f0e6572eb5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.989531] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9035bb6-d781-43d5-8763-073a2e8117ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.019916] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c903b2ba-79f6-4688-a2c5-03d755a2239b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.027418] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d18c96-2f6b-4e88-aaa4-18e3295827c7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.040611] env[61985]: DEBUG nova.compute.provider_tree [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1194.087208] env[61985]: DEBUG oslo_vmware.api [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936755, 'name': ReconfigVM_Task, 'duration_secs': 0.379789} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.087502] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Reconfigured VM instance instance-0000006f to attach disk [datastore1] volume-db955da0-6604-4fdd-85d1-b5482e8a6d9f/volume-db955da0-6604-4fdd-85d1-b5482e8a6d9f.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1194.093437] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a787a80e-916a-452c-9e71-85f1b6ad30ad {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.108612] env[61985]: DEBUG oslo_vmware.api [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1194.108612] env[61985]: value = "task-936756" [ 1194.108612] env[61985]: _type = "Task" [ 1194.108612] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.112331] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1194.112599] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-97fe4eab-c04f-40d1-b988-fcd2de9e1a17 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.121634] env[61985]: DEBUG oslo_vmware.api [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936756, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.122941] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1194.122941] env[61985]: value = "task-936757" [ 1194.122941] env[61985]: _type = "Task" [ 1194.122941] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.131547] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936757, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.256981] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1194.257188] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.257279] env[61985]: DEBUG nova.network.neutron [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1194.280412] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Getting list of instances from cluster (obj){ [ 1194.280412] env[61985]: value = "domain-c8" [ 1194.280412] env[61985]: _type = "ClusterComputeResource" [ 1194.280412] env[61985]: } {{(pid=61985) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1194.281434] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280eb869-3bd3-47ab-94f8-4f6b198279d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.299562] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Got total of 6 instances {{(pid=61985) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1194.299700] env[61985]: WARNING nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] While synchronizing instance power states, found 7 instances in the database and 6 instances on the hypervisor. [ 1194.299842] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Triggering sync for uuid 073a7668-39e6-480d-9350-835a0282b456 {{(pid=61985) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10342}} [ 1194.300042] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Triggering sync for uuid 44361f7b-4609-476f-b4a9-58a7851e6e92 {{(pid=61985) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10342}} [ 1194.300203] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Triggering sync for uuid 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 {{(pid=61985) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10342}} [ 1194.300354] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Triggering sync for uuid 45cb51fd-3876-4875-9a5d-b7760072ef55 {{(pid=61985) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10342}} [ 1194.300502] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Triggering sync for uuid 41627143-d60c-4b18-8192-ace45be71d0d {{(pid=61985) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10342}} [ 1194.300652] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Triggering sync for uuid 1b0594f2-6191-441f-b373-00b0a925d14d {{(pid=61985) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10342}} [ 1194.300838] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Triggering sync for uuid f544bc65-af75-4701-802c-f6f39f39e352 {{(pid=61985) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10342}} [ 1194.301211] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "073a7668-39e6-480d-9350-835a0282b456" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.301431] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "073a7668-39e6-480d-9350-835a0282b456" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.301694] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "44361f7b-4609-476f-b4a9-58a7851e6e92" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.301909] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.302107] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.302350] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "45cb51fd-3876-4875-9a5d-b7760072ef55" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.302846] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "41627143-d60c-4b18-8192-ace45be71d0d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.303087] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "1b0594f2-6191-441f-b373-00b0a925d14d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.303278] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "1b0594f2-6191-441f-b373-00b0a925d14d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.303510] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "f544bc65-af75-4701-802c-f6f39f39e352" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.303692] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1194.303826] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 1194.304522] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b01b368-a2dd-4865-99e5-ee861d21d09f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.307588] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28556d59-9f50-49db-b771-bbc94a1f5800 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.310297] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427eadfe-06a7-44e9-81b6-cdf73abc32f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.312626] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1194.543399] env[61985]: DEBUG nova.scheduler.client.report [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1194.619110] env[61985]: DEBUG oslo_vmware.api [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936756, 'name': ReconfigVM_Task, 'duration_secs': 0.141624} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.619440] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211507', 'volume_id': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'name': 'volume-db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41627143-d60c-4b18-8192-ace45be71d0d', 'attached_at': '', 'detached_at': '', 'volume_id': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'serial': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1194.632799] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936757, 'name': CreateSnapshot_Task, 'duration_secs': 0.472577} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.633601] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1194.634409] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5a69dd-60ef-427d-8e62-65952b898be8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.798761] env[61985]: DEBUG nova.network.neutron [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1194.816352] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.824734] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.522s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.825153] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "073a7668-39e6-480d-9350-835a0282b456" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.524s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.825364] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "1b0594f2-6191-441f-b373-00b0a925d14d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.522s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.938987] env[61985]: DEBUG nova.network.neutron [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating instance_info_cache with network_info: [{"id": "d3381564-0201-44c0-be83-9d7de95b0fce", "address": "fa:16:3e:49:33:f5", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3381564-02", "ovs_interfaceid": "d3381564-0201-44c0-be83-9d7de95b0fce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1194.941210] env[61985]: DEBUG nova.compute.manager [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1194.941723] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1194.941937] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1194.942118] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1194.942311] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1194.942465] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1194.942617] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1194.942837] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1194.942993] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1194.943180] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1194.943348] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1194.943524] env[61985]: DEBUG nova.virt.hardware [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1194.944604] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078227cd-12ca-477f-a48e-a43c2b396545 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.953995] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4451f276-ede7-48fb-a21c-2c3be29fb744 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.152667] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1195.153031] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4d187481-9e37-4f4f-9453-396d34feffa4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.163469] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1195.163469] env[61985]: value = "task-936758" [ 1195.163469] env[61985]: _type = "Task" [ 1195.163469] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.175298] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936758, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.445481] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1195.445915] env[61985]: DEBUG nova.compute.manager [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Instance network_info: |[{"id": "d3381564-0201-44c0-be83-9d7de95b0fce", "address": "fa:16:3e:49:33:f5", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3381564-02", "ovs_interfaceid": "d3381564-0201-44c0-be83-9d7de95b0fce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1195.446458] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:33:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3381564-0201-44c0-be83-9d7de95b0fce', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1195.456408] env[61985]: DEBUG oslo.service.loopingcall [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1195.456736] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1195.457053] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4aeea2e7-0aa9-49b0-897a-8d6b96a80410 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.482334] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1195.482334] env[61985]: value = "task-936759" [ 1195.482334] env[61985]: _type = "Task" [ 1195.482334] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.491318] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936759, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.554185] env[61985]: DEBUG oslo_concurrency.lockutils [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.189s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.557793] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.741s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.558062] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.558236] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1195.559482] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f2af81-aceb-498d-8190-0f024872dd13 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.570536] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4a1bef-1e7e-452b-ac46-96903a6a3ce5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.589398] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d395ce2-a91b-41c9-bf41-259bfbca8f32 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.598817] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea331a7-5f55-42fd-b3ca-ad40e29e4681 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.636759] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180660MB free_disk=43GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1195.637411] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.637411] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.660196] env[61985]: DEBUG nova.compute.manager [req-029f6d88-fda6-4e87-adb9-c7fa79f3cb79 req-c494f210-00ad-4bb0-93f7-7eda05e6fba6 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Received event network-changed-d3381564-0201-44c0-be83-9d7de95b0fce {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1195.660818] env[61985]: DEBUG nova.compute.manager [req-029f6d88-fda6-4e87-adb9-c7fa79f3cb79 req-c494f210-00ad-4bb0-93f7-7eda05e6fba6 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Refreshing instance network info cache due to event network-changed-d3381564-0201-44c0-be83-9d7de95b0fce. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1195.661336] env[61985]: DEBUG oslo_concurrency.lockutils [req-029f6d88-fda6-4e87-adb9-c7fa79f3cb79 req-c494f210-00ad-4bb0-93f7-7eda05e6fba6 service nova] Acquiring lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1195.662471] env[61985]: DEBUG oslo_concurrency.lockutils [req-029f6d88-fda6-4e87-adb9-c7fa79f3cb79 req-c494f210-00ad-4bb0-93f7-7eda05e6fba6 service nova] Acquired lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.662471] env[61985]: DEBUG nova.network.neutron [req-029f6d88-fda6-4e87-adb9-c7fa79f3cb79 req-c494f210-00ad-4bb0-93f7-7eda05e6fba6 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Refreshing network info cache for port d3381564-0201-44c0-be83-9d7de95b0fce {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1195.671813] env[61985]: DEBUG nova.objects.instance [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lazy-loading 'flavor' on Instance uuid 41627143-d60c-4b18-8192-ace45be71d0d {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.682788] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936758, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.999023] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936759, 'name': CreateVM_Task, 'duration_secs': 0.393584} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.999575] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1196.001305] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211506', 'volume_id': '8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'name': 'volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f544bc65-af75-4701-802c-f6f39f39e352', 'attached_at': '', 'detached_at': '', 'volume_id': '8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'serial': '8253a6a8-8fe6-4148-adf3-55531c7a94c3'}, 'device_type': None, 'attachment_id': '7d6ad0a0-e18d-47d1-9fab-7be7646ffef4', 'mount_device': '/dev/sda', 'disk_bus': None, 'delete_on_termination': True, 'boot_index': 0, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=61985) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1196.002303] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Root volume attach. Driver type: vmdk {{(pid=61985) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1196.003643] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e14d89-190e-4aca-a02a-ded5f3a15ba4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.017897] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d193ce62-d9bc-497e-a2c1-a4758e0fcd96 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.031126] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341ceef8-cb48-40b3-92ab-43759c565280 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.045540] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-1493fa74-3d32-4235-8527-b933567819dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.056095] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1196.056095] env[61985]: value = "task-936760" [ 1196.056095] env[61985]: _type = "Task" [ 1196.056095] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.069120] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.126255] env[61985]: INFO nova.scheduler.client.report [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleted allocation for migration 4143b7bd-7349-40ce-b676-b590deb22b76 [ 1196.180596] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936758, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.181870] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8196948b-dd83-4247-a1d8-bfaf6cd3f0e5 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.307s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.183240] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "41627143-d60c-4b18-8192-ace45be71d0d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.880s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.184517] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b4eb0b-13c5-4ed5-ab66-6f2d1f104c62 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.377610] env[61985]: DEBUG oslo_concurrency.lockutils [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "41627143-d60c-4b18-8192-ace45be71d0d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.411426] env[61985]: DEBUG nova.network.neutron [req-029f6d88-fda6-4e87-adb9-c7fa79f3cb79 req-c494f210-00ad-4bb0-93f7-7eda05e6fba6 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updated VIF entry in instance network info cache for port d3381564-0201-44c0-be83-9d7de95b0fce. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1196.411831] env[61985]: DEBUG nova.network.neutron [req-029f6d88-fda6-4e87-adb9-c7fa79f3cb79 req-c494f210-00ad-4bb0-93f7-7eda05e6fba6 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating instance_info_cache with network_info: [{"id": "d3381564-0201-44c0-be83-9d7de95b0fce", "address": "fa:16:3e:49:33:f5", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3381564-02", "ovs_interfaceid": "d3381564-0201-44c0-be83-9d7de95b0fce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.567716] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 34%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.595521] env[61985]: DEBUG oslo_concurrency.lockutils [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "44361f7b-4609-476f-b4a9-58a7851e6e92" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.635037] env[61985]: DEBUG oslo_concurrency.lockutils [None req-12adf183-d30c-48b3-8d1f-1c3da2477361 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.206s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.636155] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.334s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.637233] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7eb93a-697f-442b-83c1-e89101a2eb5a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.668395] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 073a7668-39e6-480d-9350-835a0282b456 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1196.668575] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 45cb51fd-3876-4875-9a5d-b7760072ef55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1196.668693] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 41627143-d60c-4b18-8192-ace45be71d0d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1196.668855] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 1b0594f2-6191-441f-b373-00b0a925d14d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1196.668974] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1196.669079] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 44361f7b-4609-476f-b4a9-58a7851e6e92 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1196.669185] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance f544bc65-af75-4701-802c-f6f39f39e352 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1196.669899] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1196.669899] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1196.685147] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936758, 'name': CloneVM_Task, 'duration_secs': 1.075199} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.685636] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Created linked-clone VM from snapshot [ 1196.686696] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd8f0a5-41eb-4ef3-bbb6-ad4d11abfdf7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.698671] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "41627143-d60c-4b18-8192-ace45be71d0d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.515s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.699019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.322s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.705100] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Uploading image 7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3 {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1196.734601] env[61985]: DEBUG oslo_vmware.rw_handles [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1196.734601] env[61985]: value = "vm-211509" [ 1196.734601] env[61985]: _type = "VirtualMachine" [ 1196.734601] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1196.734942] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-070c431a-abaa-4387-ab81-6b0736259a6f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.749333] env[61985]: DEBUG oslo_vmware.rw_handles [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lease: (returnval){ [ 1196.749333] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f85f3f-e4cc-7b06-4873-d6ce64d5820e" [ 1196.749333] env[61985]: _type = "HttpNfcLease" [ 1196.749333] env[61985]: } obtained for exporting VM: (result){ [ 1196.749333] env[61985]: value = "vm-211509" [ 1196.749333] env[61985]: _type = "VirtualMachine" [ 1196.749333] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1196.749723] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the lease: (returnval){ [ 1196.749723] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f85f3f-e4cc-7b06-4873-d6ce64d5820e" [ 1196.749723] env[61985]: _type = "HttpNfcLease" [ 1196.749723] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1196.762434] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1196.762434] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f85f3f-e4cc-7b06-4873-d6ce64d5820e" [ 1196.762434] env[61985]: _type = "HttpNfcLease" [ 1196.762434] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1196.794997] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa1ef65-c728-437f-bc25-a5531b7174cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.805657] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2667681-6b99-428b-9092-dedb7c8b27a9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.840882] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582e3209-b13e-4c49-b560-e884f824a76b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.852184] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b38c623-b959-476b-becd-af71bf6c82f7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.871016] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1196.914762] env[61985]: DEBUG oslo_concurrency.lockutils [req-029f6d88-fda6-4e87-adb9-c7fa79f3cb79 req-c494f210-00ad-4bb0-93f7-7eda05e6fba6 service nova] Releasing lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1197.070150] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 45%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.154526] env[61985]: INFO nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] During sync_power_state the instance has a pending task (deleting). Skip. [ 1197.154797] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.519s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.155086] env[61985]: DEBUG oslo_concurrency.lockutils [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.560s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.155325] env[61985]: DEBUG oslo_concurrency.lockutils [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "44361f7b-4609-476f-b4a9-58a7851e6e92-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.155536] env[61985]: DEBUG oslo_concurrency.lockutils [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.155790] env[61985]: DEBUG oslo_concurrency.lockutils [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.158644] env[61985]: INFO nova.compute.manager [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Terminating instance [ 1197.160731] env[61985]: DEBUG nova.compute.manager [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1197.160927] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1197.161795] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a38c68-cb37-4b9e-95aa-f81eee091684 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.173125] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1197.173437] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-559511fb-ca30-44b2-b849-447f6c62c10e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.182347] env[61985]: DEBUG oslo_vmware.api [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1197.182347] env[61985]: value = "task-936762" [ 1197.182347] env[61985]: _type = "Task" [ 1197.182347] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.195204] env[61985]: DEBUG oslo_vmware.api [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936762, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.211332] env[61985]: INFO nova.compute.manager [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Detaching volume db955da0-6604-4fdd-85d1-b5482e8a6d9f [ 1197.256475] env[61985]: INFO nova.virt.block_device [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Attempting to driver detach volume db955da0-6604-4fdd-85d1-b5482e8a6d9f from mountpoint /dev/sdb [ 1197.256720] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1197.256947] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211507', 'volume_id': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'name': 'volume-db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41627143-d60c-4b18-8192-ace45be71d0d', 'attached_at': '', 'detached_at': '', 'volume_id': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'serial': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1197.258122] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1b24ea-c72a-4d8f-ab8e-632348916053 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.262868] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1197.262868] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f85f3f-e4cc-7b06-4873-d6ce64d5820e" [ 1197.262868] env[61985]: _type = "HttpNfcLease" [ 1197.262868] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1197.263637] env[61985]: DEBUG oslo_vmware.rw_handles [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1197.263637] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f85f3f-e4cc-7b06-4873-d6ce64d5820e" [ 1197.263637] env[61985]: _type = "HttpNfcLease" [ 1197.263637] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1197.264440] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c5f5b6-af7d-4d5b-92cb-ba3887e8d755 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.289061] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877f0a27-eb3e-4330-a854-4eb67eb8d8be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.292303] env[61985]: DEBUG oslo_vmware.rw_handles [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d68a8-7dff-84f0-af65-c6e77dcfc258/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1197.292491] env[61985]: DEBUG oslo_vmware.rw_handles [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d68a8-7dff-84f0-af65-c6e77dcfc258/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1197.367973] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e329ddd-86d5-4535-b32e-aebc85bcccd7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.374087] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1197.396576] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a093100-fa6c-4ad1-9029-60c45344b631 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.415229] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] The volume has not been displaced from its original location: [datastore1] volume-db955da0-6604-4fdd-85d1-b5482e8a6d9f/volume-db955da0-6604-4fdd-85d1-b5482e8a6d9f.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1197.420687] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1197.421075] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-322cbdf2-5dd8-4171-af5a-d41f86f1f4b2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.443668] env[61985]: DEBUG oslo_vmware.api [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1197.443668] env[61985]: value = "task-936763" [ 1197.443668] env[61985]: _type = "Task" [ 1197.443668] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.459193] env[61985]: DEBUG oslo_vmware.api [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936763, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.496122] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a1256fd8-82b8-48f2-b949-50e6af30b834 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.572682] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 58%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.699097] env[61985]: DEBUG oslo_vmware.api [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936762, 'name': PowerOffVM_Task, 'duration_secs': 0.438313} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.699415] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1197.699616] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1197.699963] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a162e6ab-0b08-4525-a64b-10457a7b82be {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.798431] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1197.798694] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1197.799075] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleting the datastore file [datastore1] 44361f7b-4609-476f-b4a9-58a7851e6e92 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1197.799439] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb756419-56c3-4b74-af4b-39f695e16828 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.810883] env[61985]: DEBUG oslo_vmware.api [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1197.810883] env[61985]: value = "task-936765" [ 1197.810883] env[61985]: _type = "Task" [ 1197.810883] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.822496] env[61985]: DEBUG oslo_vmware.api [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.896926] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1197.897173] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.260s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.956934] env[61985]: DEBUG oslo_vmware.api [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936763, 'name': ReconfigVM_Task, 'duration_secs': 0.323724} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.957298] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1197.962400] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c3a9d0e-99a5-4e6a-b0c9-3d9aea445f9f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.981703] env[61985]: DEBUG oslo_vmware.api [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1197.981703] env[61985]: value = "task-936766" [ 1197.981703] env[61985]: _type = "Task" [ 1197.981703] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.994178] env[61985]: DEBUG oslo_vmware.api [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936766, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.075158] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 71%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.324032] env[61985]: DEBUG oslo_vmware.api [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936765, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.324268} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.324330] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1198.324527] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1198.324718] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1198.324901] env[61985]: INFO nova.compute.manager [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1198.325182] env[61985]: DEBUG oslo.service.loopingcall [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.325396] env[61985]: DEBUG nova.compute.manager [-] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1198.325509] env[61985]: DEBUG nova.network.neutron [-] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1198.498471] env[61985]: DEBUG oslo_vmware.api [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936766, 'name': ReconfigVM_Task, 'duration_secs': 0.201076} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.498823] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211507', 'volume_id': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'name': 'volume-db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41627143-d60c-4b18-8192-ace45be71d0d', 'attached_at': '', 'detached_at': '', 'volume_id': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f', 'serial': 'db955da0-6604-4fdd-85d1-b5482e8a6d9f'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1198.573368] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 84%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.921897] env[61985]: DEBUG nova.compute.manager [req-2fc488b9-120c-43cf-85f9-ba240b8b37cf req-8427913d-75c0-41f6-af2b-574479066280 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Received event network-vif-deleted-e35f2970-0b73-4b31-925f-56ae8e0e7f39 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1198.922723] env[61985]: INFO nova.compute.manager [req-2fc488b9-120c-43cf-85f9-ba240b8b37cf req-8427913d-75c0-41f6-af2b-574479066280 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Neutron deleted interface e35f2970-0b73-4b31-925f-56ae8e0e7f39; detaching it from the instance and deleting it from the info cache [ 1198.923069] env[61985]: DEBUG nova.network.neutron [req-2fc488b9-120c-43cf-85f9-ba240b8b37cf req-8427913d-75c0-41f6-af2b-574479066280 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.055357] env[61985]: DEBUG nova.objects.instance [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lazy-loading 'flavor' on Instance uuid 41627143-d60c-4b18-8192-ace45be71d0d {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.072522] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 97%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.389440] env[61985]: DEBUG nova.network.neutron [-] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.427031] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36b812a4-6c88-4663-80b8-c66cb916467a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.437312] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2697ae52-b228-4d0e-a99f-2c25d97f29c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.470728] env[61985]: DEBUG nova.compute.manager [req-2fc488b9-120c-43cf-85f9-ba240b8b37cf req-8427913d-75c0-41f6-af2b-574479066280 service nova] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Detach interface failed, port_id=e35f2970-0b73-4b31-925f-56ae8e0e7f39, reason: Instance 44361f7b-4609-476f-b4a9-58a7851e6e92 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1199.572125] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 98%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.892049] env[61985]: INFO nova.compute.manager [-] [instance: 44361f7b-4609-476f-b4a9-58a7851e6e92] Took 1.57 seconds to deallocate network for instance. [ 1200.063291] env[61985]: DEBUG oslo_concurrency.lockutils [None req-469b85a1-cb52-411d-a85f-b430fa301f29 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.364s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.074617] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 98%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.399375] env[61985]: DEBUG oslo_concurrency.lockutils [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.399558] env[61985]: DEBUG oslo_concurrency.lockutils [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.399813] env[61985]: DEBUG nova.objects.instance [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'resources' on Instance uuid 44361f7b-4609-476f-b4a9-58a7851e6e92 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1200.576487] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 98%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.767104] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "41627143-d60c-4b18-8192-ace45be71d0d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.767407] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.767634] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "41627143-d60c-4b18-8192-ace45be71d0d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.767918] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.768126] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.770228] env[61985]: INFO nova.compute.manager [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Terminating instance [ 1200.772035] env[61985]: DEBUG nova.compute.manager [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1200.772233] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1200.773071] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b58319a-42f8-498e-b044-eff842090937 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.781582] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1200.782181] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-351d7e28-f140-4335-8e34-f4d0033a4933 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.791892] env[61985]: DEBUG oslo_vmware.api [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1200.791892] env[61985]: value = "task-936767" [ 1200.791892] env[61985]: _type = "Task" [ 1200.791892] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.801886] env[61985]: DEBUG oslo_vmware.api [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936767, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.009508] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7acef8b0-f888-421b-be04-981885097da2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.018092] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db84a20-1448-43fd-b23d-3646500cfc8f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.048483] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d200a15-3bac-41c2-9e05-9b3f9fcc8c04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.056693] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bec76da-8e37-471a-94a9-4f1edda71e78 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.070939] env[61985]: DEBUG nova.compute.provider_tree [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1201.080558] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 98%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.303197] env[61985]: DEBUG oslo_vmware.api [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936767, 'name': PowerOffVM_Task, 'duration_secs': 0.198155} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.303513] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1201.303669] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1201.303924] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c208078b-af0c-427f-9331-3c175b04734d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.550082] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1201.550347] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1201.550641] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleting the datastore file [datastore1] 41627143-d60c-4b18-8192-ace45be71d0d {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1201.550966] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29dc479a-6d38-42b8-aaac-f150aa4dffd5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.559565] env[61985]: DEBUG oslo_vmware.api [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for the task: (returnval){ [ 1201.559565] env[61985]: value = "task-936769" [ 1201.559565] env[61985]: _type = "Task" [ 1201.559565] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.568089] env[61985]: DEBUG oslo_vmware.api [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936769, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.576508] env[61985]: DEBUG nova.scheduler.client.report [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1201.582469] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task} progress is 98%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.071245] env[61985]: DEBUG oslo_vmware.api [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Task: {'id': task-936769, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138953} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.071618] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1202.071618] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1202.071805] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1202.071993] env[61985]: INFO nova.compute.manager [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Took 1.30 seconds to destroy the instance on the hypervisor. [ 1202.072268] env[61985]: DEBUG oslo.service.loopingcall [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1202.072527] env[61985]: DEBUG nova.compute.manager [-] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1202.072623] env[61985]: DEBUG nova.network.neutron [-] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1202.082800] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936760, 'name': RelocateVM_Task, 'duration_secs': 5.99093} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.083480] env[61985]: DEBUG oslo_concurrency.lockutils [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.684s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.085338] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1202.085559] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211506', 'volume_id': '8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'name': 'volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f544bc65-af75-4701-802c-f6f39f39e352', 'attached_at': '', 'detached_at': '', 'volume_id': '8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'serial': '8253a6a8-8fe6-4148-adf3-55531c7a94c3'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1202.086546] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544aba6f-7936-4eb7-a623-1fd44e0e8b1c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.105741] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129c2be5-ce22-4700-8646-efc4ad157ca3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.111302] env[61985]: INFO nova.scheduler.client.report [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleted allocations for instance 44361f7b-4609-476f-b4a9-58a7851e6e92 [ 1202.133562] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3/volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1202.134021] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab5ee096-b3a8-4863-b8f7-926e96b7ac27 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.156485] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1202.156485] env[61985]: value = "task-936770" [ 1202.156485] env[61985]: _type = "Task" [ 1202.156485] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.166440] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936770, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.628148] env[61985]: DEBUG nova.compute.manager [req-0cecf00f-3aa8-4af5-8e23-cf9dcdec48cc req-5fa52dea-1dea-421c-aa12-aeb2fb26cb0a service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Received event network-vif-deleted-bb5ccfcb-7752-47cd-b569-3c2861ea2a73 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1202.628350] env[61985]: INFO nova.compute.manager [req-0cecf00f-3aa8-4af5-8e23-cf9dcdec48cc req-5fa52dea-1dea-421c-aa12-aeb2fb26cb0a service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Neutron deleted interface bb5ccfcb-7752-47cd-b569-3c2861ea2a73; detaching it from the instance and deleting it from the info cache [ 1202.628605] env[61985]: DEBUG nova.network.neutron [req-0cecf00f-3aa8-4af5-8e23-cf9dcdec48cc req-5fa52dea-1dea-421c-aa12-aeb2fb26cb0a service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1202.637412] env[61985]: DEBUG oslo_concurrency.lockutils [None req-254491cb-758f-4194-95ad-1ca0c8c87269 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "44361f7b-4609-476f-b4a9-58a7851e6e92" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.482s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.667748] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936770, 'name': ReconfigVM_Task, 'duration_secs': 0.414619} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.667945] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Reconfigured VM instance instance-00000072 to attach disk [datastore1] volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3/volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.673068] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-081b1a43-af5d-46be-9af9-1c4e2c411c7c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.690184] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1202.690184] env[61985]: value = "task-936771" [ 1202.690184] env[61985]: _type = "Task" [ 1202.690184] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.700059] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936771, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.071263] env[61985]: DEBUG nova.network.neutron [-] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1203.131846] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c0312d0-b082-4ba3-891f-54a174c129cd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.144078] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cce901-a52e-4b39-b947-0a70be330ff2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.179654] env[61985]: DEBUG nova.compute.manager [req-0cecf00f-3aa8-4af5-8e23-cf9dcdec48cc req-5fa52dea-1dea-421c-aa12-aeb2fb26cb0a service nova] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Detach interface failed, port_id=bb5ccfcb-7752-47cd-b569-3c2861ea2a73, reason: Instance 41627143-d60c-4b18-8192-ace45be71d0d could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1203.203469] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936771, 'name': ReconfigVM_Task, 'duration_secs': 0.148393} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.203902] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211506', 'volume_id': '8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'name': 'volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f544bc65-af75-4701-802c-f6f39f39e352', 'attached_at': '', 'detached_at': '', 'volume_id': '8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'serial': '8253a6a8-8fe6-4148-adf3-55531c7a94c3'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1203.204432] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7666e80c-f5a2-467b-80c3-fd3645ed18b9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.213511] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1203.213511] env[61985]: value = "task-936772" [ 1203.213511] env[61985]: _type = "Task" [ 1203.213511] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.224097] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936772, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.573667] env[61985]: INFO nova.compute.manager [-] [instance: 41627143-d60c-4b18-8192-ace45be71d0d] Took 1.50 seconds to deallocate network for instance. [ 1203.724681] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936772, 'name': Rename_Task, 'duration_secs': 0.145697} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.724963] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1203.725246] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8296fe7-bd8a-4731-8f5e-d8efe4a4c861 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.733828] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1203.733828] env[61985]: value = "task-936773" [ 1203.733828] env[61985]: _type = "Task" [ 1203.733828] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.743482] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936773, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.913729] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.913993] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.080705] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.081144] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.081421] env[61985]: DEBUG nova.objects.instance [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lazy-loading 'resources' on Instance uuid 41627143-d60c-4b18-8192-ace45be71d0d {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1204.244463] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936773, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.417156] env[61985]: DEBUG nova.compute.manager [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1204.687204] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459c282e-a199-40e2-9d8f-a5c0e1416bfd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.695977] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1fde77f-d03c-47dc-9cf9-a96a09c96814 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.728149] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef186ba-77dd-4f74-bc59-95f478cd476c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.739867] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3944e89-3697-4ed3-b021-4cfb55ce27a9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.749669] env[61985]: DEBUG oslo_vmware.api [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936773, 'name': PowerOnVM_Task, 'duration_secs': 0.734199} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.758050] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1204.758341] env[61985]: INFO nova.compute.manager [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Took 9.82 seconds to spawn the instance on the hypervisor. [ 1204.758567] env[61985]: DEBUG nova.compute.manager [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1204.759146] env[61985]: DEBUG nova.compute.provider_tree [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1204.760923] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e411c7-0556-470b-aceb-978c1124e69a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.939177] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.265708] env[61985]: DEBUG nova.scheduler.client.report [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1205.278556] env[61985]: INFO nova.compute.manager [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Took 16.52 seconds to build instance. [ 1205.716431] env[61985]: DEBUG nova.compute.manager [req-ddc5d124-69ee-42b1-ba0b-a9ba707cc584 req-6acb90b9-bd39-4b49-a174-217766f35c13 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Received event network-changed-a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1205.716629] env[61985]: DEBUG nova.compute.manager [req-ddc5d124-69ee-42b1-ba0b-a9ba707cc584 req-6acb90b9-bd39-4b49-a174-217766f35c13 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Refreshing instance network info cache due to event network-changed-a5e0f361-c876-4212-8c4a-dfdfb17d5d4f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1205.716959] env[61985]: DEBUG oslo_concurrency.lockutils [req-ddc5d124-69ee-42b1-ba0b-a9ba707cc584 req-6acb90b9-bd39-4b49-a174-217766f35c13 service nova] Acquiring lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1205.717288] env[61985]: DEBUG oslo_concurrency.lockutils [req-ddc5d124-69ee-42b1-ba0b-a9ba707cc584 req-6acb90b9-bd39-4b49-a174-217766f35c13 service nova] Acquired lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.717480] env[61985]: DEBUG nova.network.neutron [req-ddc5d124-69ee-42b1-ba0b-a9ba707cc584 req-6acb90b9-bd39-4b49-a174-217766f35c13 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Refreshing network info cache for port a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1205.770401] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.689s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.773499] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.834s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.778137] env[61985]: INFO nova.compute.claims [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1205.780329] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6d76d2d8-b879-441c-aabc-90c6361ed166 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "f544bc65-af75-4701-802c-f6f39f39e352" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.028s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.780665] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "f544bc65-af75-4701-802c-f6f39f39e352" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 11.477s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.780890] env[61985]: INFO nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: f544bc65-af75-4701-802c-f6f39f39e352] During sync_power_state the instance has a pending task (block_device_mapping). Skip. [ 1205.781115] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "f544bc65-af75-4701-802c-f6f39f39e352" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.831802] env[61985]: INFO nova.scheduler.client.report [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Deleted allocations for instance 41627143-d60c-4b18-8192-ace45be71d0d [ 1206.048551] env[61985]: DEBUG oslo_vmware.rw_handles [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d68a8-7dff-84f0-af65-c6e77dcfc258/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1206.049783] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51702d6-2c90-4672-b872-f4af0e653fd1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.057121] env[61985]: DEBUG oslo_vmware.rw_handles [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d68a8-7dff-84f0-af65-c6e77dcfc258/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1206.057328] env[61985]: ERROR oslo_vmware.rw_handles [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d68a8-7dff-84f0-af65-c6e77dcfc258/disk-0.vmdk due to incomplete transfer. [ 1206.057572] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-105ae2d2-353e-40be-a84e-4bff98ccd4f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.066544] env[61985]: DEBUG oslo_vmware.rw_handles [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524d68a8-7dff-84f0-af65-c6e77dcfc258/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1206.066763] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Uploaded image 7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3 to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1206.069302] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1206.069569] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8991f529-cbf0-4aba-844d-16ff7afe7e94 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.076777] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1206.076777] env[61985]: value = "task-936774" [ 1206.076777] env[61985]: _type = "Task" [ 1206.076777] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.085281] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936774, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.342258] env[61985]: DEBUG oslo_concurrency.lockutils [None req-b40cfd22-54f9-457c-843b-e031e209c1f4 tempest-AttachVolumeNegativeTest-1684227758 tempest-AttachVolumeNegativeTest-1684227758-project-member] Lock "41627143-d60c-4b18-8192-ace45be71d0d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.574s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.587396] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936774, 'name': Destroy_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.640225] env[61985]: DEBUG nova.network.neutron [req-ddc5d124-69ee-42b1-ba0b-a9ba707cc584 req-6acb90b9-bd39-4b49-a174-217766f35c13 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Updated VIF entry in instance network info cache for port a5e0f361-c876-4212-8c4a-dfdfb17d5d4f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1206.640643] env[61985]: DEBUG nova.network.neutron [req-ddc5d124-69ee-42b1-ba0b-a9ba707cc584 req-6acb90b9-bd39-4b49-a174-217766f35c13 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Updating instance_info_cache with network_info: [{"id": "a5e0f361-c876-4212-8c4a-dfdfb17d5d4f", "address": "fa:16:3e:3a:9a:10", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5e0f361-c8", "ovs_interfaceid": "a5e0f361-c876-4212-8c4a-dfdfb17d5d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.894550] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bcf6789-5274-4f5a-b5b3-9ef23a49d0bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.903300] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46278cfb-8b32-45cd-9a53-17bba4a1424b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.934052] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2ed22d-3313-444c-a146-fd69a4e32868 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.942967] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b5cc48f-90fd-4385-b5f7-c9ca6112e2ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.962714] env[61985]: DEBUG nova.compute.provider_tree [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1207.089856] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936774, 'name': Destroy_Task, 'duration_secs': 0.530839} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.089856] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Destroyed the VM [ 1207.089856] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1207.089856] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f32d8d3f-d075-40da-bdea-83bec7959388 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.097734] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1207.097734] env[61985]: value = "task-936775" [ 1207.097734] env[61985]: _type = "Task" [ 1207.097734] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.107560] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936775, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.143583] env[61985]: DEBUG oslo_concurrency.lockutils [req-ddc5d124-69ee-42b1-ba0b-a9ba707cc584 req-6acb90b9-bd39-4b49-a174-217766f35c13 service nova] Releasing lock "refresh_cache-073a7668-39e6-480d-9350-835a0282b456" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1207.223122] env[61985]: DEBUG nova.compute.manager [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Stashing vm_state: active {{(pid=61985) _prep_resize /opt/stack/nova/nova/compute/manager.py:5630}} [ 1207.467055] env[61985]: DEBUG nova.scheduler.client.report [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1207.608264] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936775, 'name': RemoveSnapshot_Task, 'duration_secs': 0.338168} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.608556] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1207.608859] env[61985]: DEBUG nova.compute.manager [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1207.609618] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47e900c-869c-45ab-8559-1f58f94ab492 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.740339] env[61985]: DEBUG nova.compute.manager [req-a6f2ae3c-609d-49cd-9653-1b94807c79ee req-6133d80d-f8a6-4aa1-95a4-ea298c2c6769 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Received event network-changed-d3381564-0201-44c0-be83-9d7de95b0fce {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1207.740559] env[61985]: DEBUG nova.compute.manager [req-a6f2ae3c-609d-49cd-9653-1b94807c79ee req-6133d80d-f8a6-4aa1-95a4-ea298c2c6769 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Refreshing instance network info cache due to event network-changed-d3381564-0201-44c0-be83-9d7de95b0fce. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1207.740782] env[61985]: DEBUG oslo_concurrency.lockutils [req-a6f2ae3c-609d-49cd-9653-1b94807c79ee req-6133d80d-f8a6-4aa1-95a4-ea298c2c6769 service nova] Acquiring lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1207.740934] env[61985]: DEBUG oslo_concurrency.lockutils [req-a6f2ae3c-609d-49cd-9653-1b94807c79ee req-6133d80d-f8a6-4aa1-95a4-ea298c2c6769 service nova] Acquired lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.741159] env[61985]: DEBUG nova.network.neutron [req-a6f2ae3c-609d-49cd-9653-1b94807c79ee req-6133d80d-f8a6-4aa1-95a4-ea298c2c6769 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Refreshing network info cache for port d3381564-0201-44c0-be83-9d7de95b0fce {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1207.743715] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.972447] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.199s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.972968] env[61985]: DEBUG nova.compute.manager [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1207.975822] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.233s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.122257] env[61985]: INFO nova.compute.manager [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Shelve offloading [ 1208.123848] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1208.124119] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f88ddf8-2e86-4d1c-bb43-238f20ff5792 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.134086] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1208.134086] env[61985]: value = "task-936777" [ 1208.134086] env[61985]: _type = "Task" [ 1208.134086] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.144765] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1208.144993] env[61985]: DEBUG nova.compute.manager [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1208.145827] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e561d39-f09d-46c6-bef3-d64f8ed74291 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.152822] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1208.153035] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.153351] env[61985]: DEBUG nova.network.neutron [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1208.477132] env[61985]: DEBUG nova.network.neutron [req-a6f2ae3c-609d-49cd-9653-1b94807c79ee req-6133d80d-f8a6-4aa1-95a4-ea298c2c6769 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updated VIF entry in instance network info cache for port d3381564-0201-44c0-be83-9d7de95b0fce. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1208.477531] env[61985]: DEBUG nova.network.neutron [req-a6f2ae3c-609d-49cd-9653-1b94807c79ee req-6133d80d-f8a6-4aa1-95a4-ea298c2c6769 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating instance_info_cache with network_info: [{"id": "d3381564-0201-44c0-be83-9d7de95b0fce", "address": "fa:16:3e:49:33:f5", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3381564-02", "ovs_interfaceid": "d3381564-0201-44c0-be83-9d7de95b0fce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1208.479760] env[61985]: DEBUG nova.compute.utils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1208.484529] env[61985]: INFO nova.compute.claims [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1208.488290] env[61985]: DEBUG nova.compute.manager [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1208.488464] env[61985]: DEBUG nova.network.neutron [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1208.526864] env[61985]: DEBUG nova.policy [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '905fa37e1f784a56b3edb9818b03d685', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23a4809e353a4ecaa1bd53612fb4afc5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1208.812187] env[61985]: DEBUG nova.network.neutron [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Successfully created port: 380110e4-5fc3-4dbf-b9ba-c94f348c52d2 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1208.882167] env[61985]: DEBUG nova.network.neutron [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updating instance_info_cache with network_info: [{"id": "eedcafdd-f3f2-4705-a528-258bb48b4574", "address": "fa:16:3e:49:21:b1", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeedcafdd-f3", "ovs_interfaceid": "eedcafdd-f3f2-4705-a528-258bb48b4574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1208.989165] env[61985]: DEBUG oslo_concurrency.lockutils [req-a6f2ae3c-609d-49cd-9653-1b94807c79ee req-6133d80d-f8a6-4aa1-95a4-ea298c2c6769 service nova] Releasing lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.989799] env[61985]: DEBUG nova.compute.manager [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1208.993604] env[61985]: INFO nova.compute.resource_tracker [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating resource usage from migration ef004837-0950-4c94-a03b-965efd491213 [ 1209.101540] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4682b18e-d063-436c-b940-e2e21326406c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.110387] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840194d7-1882-4194-8c50-208cf7576eef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.141603] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6f41c2-0a45-40df-a37e-87bcedb10305 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.150122] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a198f29-4803-4813-a9b8-7d7eeff98812 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.163441] env[61985]: DEBUG nova.compute.provider_tree [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.385283] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1209.661157] env[61985]: DEBUG nova.compute.manager [req-43f7c673-8cd6-48d3-88ca-940f2e0b8a66 req-30f9d34e-7380-4f6f-9d3b-f5960254c499 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received event network-vif-unplugged-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1209.661429] env[61985]: DEBUG oslo_concurrency.lockutils [req-43f7c673-8cd6-48d3-88ca-940f2e0b8a66 req-30f9d34e-7380-4f6f-9d3b-f5960254c499 service nova] Acquiring lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.661616] env[61985]: DEBUG oslo_concurrency.lockutils [req-43f7c673-8cd6-48d3-88ca-940f2e0b8a66 req-30f9d34e-7380-4f6f-9d3b-f5960254c499 service nova] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.661789] env[61985]: DEBUG oslo_concurrency.lockutils [req-43f7c673-8cd6-48d3-88ca-940f2e0b8a66 req-30f9d34e-7380-4f6f-9d3b-f5960254c499 service nova] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.661965] env[61985]: DEBUG nova.compute.manager [req-43f7c673-8cd6-48d3-88ca-940f2e0b8a66 req-30f9d34e-7380-4f6f-9d3b-f5960254c499 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] No waiting events found dispatching network-vif-unplugged-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1209.662185] env[61985]: WARNING nova.compute.manager [req-43f7c673-8cd6-48d3-88ca-940f2e0b8a66 req-30f9d34e-7380-4f6f-9d3b-f5960254c499 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received unexpected event network-vif-unplugged-eedcafdd-f3f2-4705-a528-258bb48b4574 for instance with vm_state shelved and task_state shelving_offloading. [ 1209.665754] env[61985]: DEBUG nova.scheduler.client.report [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1209.711908] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1209.712838] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0c3181-08a8-474d-9cb8-da7d8a1ea24f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.721441] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1209.721673] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cabeaba2-7865-4b46-b749-ffcd9c19485a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.821849] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1209.822147] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1209.822388] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleting the datastore file [datastore1] 45cb51fd-3876-4875-9a5d-b7760072ef55 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1209.822689] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a5ccb8e8-8024-4efc-a0f9-4859775f9775 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.831029] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1209.831029] env[61985]: value = "task-936779" [ 1209.831029] env[61985]: _type = "Task" [ 1209.831029] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.839231] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.003205] env[61985]: DEBUG nova.compute.manager [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1210.029134] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1210.029432] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1210.029641] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1210.029849] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1210.030578] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1210.030824] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1210.031106] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1210.031325] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1210.031569] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1210.031769] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1210.031992] env[61985]: DEBUG nova.virt.hardware [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1210.032849] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3304b453-dc12-4025-8c92-e308669aee38 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.041238] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84c4baa-931e-4d77-bb24-c31aa74dfd1d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.172016] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.194s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.172016] env[61985]: INFO nova.compute.manager [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Migrating [ 1210.196263] env[61985]: DEBUG nova.compute.manager [req-e2a2b695-8140-488a-a8c8-01ae1cad7275 req-5db11957-ab49-473a-b105-bc2dee583533 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Received event network-vif-plugged-380110e4-5fc3-4dbf-b9ba-c94f348c52d2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1210.196263] env[61985]: DEBUG oslo_concurrency.lockutils [req-e2a2b695-8140-488a-a8c8-01ae1cad7275 req-5db11957-ab49-473a-b105-bc2dee583533 service nova] Acquiring lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.196263] env[61985]: DEBUG oslo_concurrency.lockutils [req-e2a2b695-8140-488a-a8c8-01ae1cad7275 req-5db11957-ab49-473a-b105-bc2dee583533 service nova] Lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.196263] env[61985]: DEBUG oslo_concurrency.lockutils [req-e2a2b695-8140-488a-a8c8-01ae1cad7275 req-5db11957-ab49-473a-b105-bc2dee583533 service nova] Lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.196263] env[61985]: DEBUG nova.compute.manager [req-e2a2b695-8140-488a-a8c8-01ae1cad7275 req-5db11957-ab49-473a-b105-bc2dee583533 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] No waiting events found dispatching network-vif-plugged-380110e4-5fc3-4dbf-b9ba-c94f348c52d2 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1210.196263] env[61985]: WARNING nova.compute.manager [req-e2a2b695-8140-488a-a8c8-01ae1cad7275 req-5db11957-ab49-473a-b105-bc2dee583533 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Received unexpected event network-vif-plugged-380110e4-5fc3-4dbf-b9ba-c94f348c52d2 for instance with vm_state building and task_state spawning. [ 1210.273675] env[61985]: DEBUG nova.network.neutron [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Successfully updated port: 380110e4-5fc3-4dbf-b9ba-c94f348c52d2 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1210.342508] env[61985]: DEBUG oslo_vmware.api [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936779, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142784} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.342799] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1210.342991] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1210.343216] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1210.363263] env[61985]: INFO nova.scheduler.client.report [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleted allocations for instance 45cb51fd-3876-4875-9a5d-b7760072ef55 [ 1210.691034] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.691411] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquired lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.691800] env[61985]: DEBUG nova.network.neutron [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1210.777047] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.777231] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.777366] env[61985]: DEBUG nova.network.neutron [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1210.867914] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.868213] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.868445] env[61985]: DEBUG nova.objects.instance [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'resources' on Instance uuid 45cb51fd-3876-4875-9a5d-b7760072ef55 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1211.307035] env[61985]: DEBUG nova.network.neutron [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1211.371860] env[61985]: DEBUG nova.objects.instance [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'numa_topology' on Instance uuid 45cb51fd-3876-4875-9a5d-b7760072ef55 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1211.410012] env[61985]: DEBUG nova.network.neutron [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating instance_info_cache with network_info: [{"id": "d3381564-0201-44c0-be83-9d7de95b0fce", "address": "fa:16:3e:49:33:f5", "network": {"id": "3cfbbfa5-703f-4314-8e11-9ce34250383a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1650934949-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bdb5f76b9b24a7ba4800a88482847da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3381564-02", "ovs_interfaceid": "d3381564-0201-44c0-be83-9d7de95b0fce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.436392] env[61985]: DEBUG nova.network.neutron [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance_info_cache with network_info: [{"id": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "address": "fa:16:3e:56:1d:3d", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380110e4-5f", "ovs_interfaceid": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.702731] env[61985]: DEBUG nova.compute.manager [req-39267dfa-9dc3-42c3-b056-dc19d159e8f3 req-0077cca4-8ade-4cfc-9195-36e3b4608d9e service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received event network-changed-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1211.703103] env[61985]: DEBUG nova.compute.manager [req-39267dfa-9dc3-42c3-b056-dc19d159e8f3 req-0077cca4-8ade-4cfc-9195-36e3b4608d9e service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Refreshing instance network info cache due to event network-changed-eedcafdd-f3f2-4705-a528-258bb48b4574. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1211.703172] env[61985]: DEBUG oslo_concurrency.lockutils [req-39267dfa-9dc3-42c3-b056-dc19d159e8f3 req-0077cca4-8ade-4cfc-9195-36e3b4608d9e service nova] Acquiring lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1211.703315] env[61985]: DEBUG oslo_concurrency.lockutils [req-39267dfa-9dc3-42c3-b056-dc19d159e8f3 req-0077cca4-8ade-4cfc-9195-36e3b4608d9e service nova] Acquired lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.703481] env[61985]: DEBUG nova.network.neutron [req-39267dfa-9dc3-42c3-b056-dc19d159e8f3 req-0077cca4-8ade-4cfc-9195-36e3b4608d9e service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Refreshing network info cache for port eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1211.877090] env[61985]: DEBUG nova.objects.base [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Object Instance<45cb51fd-3876-4875-9a5d-b7760072ef55> lazy-loaded attributes: resources,numa_topology {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1211.914568] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Releasing lock "refresh_cache-f544bc65-af75-4701-802c-f6f39f39e352" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1211.939777] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1211.940129] env[61985]: DEBUG nova.compute.manager [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Instance network_info: |[{"id": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "address": "fa:16:3e:56:1d:3d", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380110e4-5f", "ovs_interfaceid": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1211.940763] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:1d:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '380110e4-5fc3-4dbf-b9ba-c94f348c52d2', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1211.948590] env[61985]: DEBUG oslo.service.loopingcall [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1211.951378] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1211.951792] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b754152-0b50-4fb1-94f4-ffa757b6c6fe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.977064] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1211.977064] env[61985]: value = "task-936781" [ 1211.977064] env[61985]: _type = "Task" [ 1211.977064] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.983778] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9224ed1c-730a-4879-88f5-4d781271235e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.989306] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936781, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.994624] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15de19af-e881-411e-a978-343fa2af4484 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.029822] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a733d3fa-697f-43ce-9bac-68c1b9058d2d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.038493] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11b8860-2212-400c-981c-510ca15907c8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.054814] env[61985]: DEBUG nova.compute.provider_tree [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1212.223622] env[61985]: DEBUG nova.compute.manager [req-7de83242-4317-4bda-8bdc-4880294bb35c req-89f5cefd-0b68-4cfd-b16e-94fdcafb6410 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Received event network-changed-380110e4-5fc3-4dbf-b9ba-c94f348c52d2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1212.223850] env[61985]: DEBUG nova.compute.manager [req-7de83242-4317-4bda-8bdc-4880294bb35c req-89f5cefd-0b68-4cfd-b16e-94fdcafb6410 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Refreshing instance network info cache due to event network-changed-380110e4-5fc3-4dbf-b9ba-c94f348c52d2. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1212.224084] env[61985]: DEBUG oslo_concurrency.lockutils [req-7de83242-4317-4bda-8bdc-4880294bb35c req-89f5cefd-0b68-4cfd-b16e-94fdcafb6410 service nova] Acquiring lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.224240] env[61985]: DEBUG oslo_concurrency.lockutils [req-7de83242-4317-4bda-8bdc-4880294bb35c req-89f5cefd-0b68-4cfd-b16e-94fdcafb6410 service nova] Acquired lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.224411] env[61985]: DEBUG nova.network.neutron [req-7de83242-4317-4bda-8bdc-4880294bb35c req-89f5cefd-0b68-4cfd-b16e-94fdcafb6410 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Refreshing network info cache for port 380110e4-5fc3-4dbf-b9ba-c94f348c52d2 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1212.431306] env[61985]: DEBUG nova.network.neutron [req-39267dfa-9dc3-42c3-b056-dc19d159e8f3 req-0077cca4-8ade-4cfc-9195-36e3b4608d9e service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updated VIF entry in instance network info cache for port eedcafdd-f3f2-4705-a528-258bb48b4574. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1212.431651] env[61985]: DEBUG nova.network.neutron [req-39267dfa-9dc3-42c3-b056-dc19d159e8f3 req-0077cca4-8ade-4cfc-9195-36e3b4608d9e service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updating instance_info_cache with network_info: [{"id": "eedcafdd-f3f2-4705-a528-258bb48b4574", "address": "fa:16:3e:49:21:b1", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapeedcafdd-f3", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.487263] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936781, 'name': CreateVM_Task, 'duration_secs': 0.344264} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.487575] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1212.488157] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.488326] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.488650] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1212.489083] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91b011a3-ff6d-4558-ab9f-46881db7234b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.493939] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1212.493939] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521e79ef-b336-3859-5b0a-c2f08c1eb402" [ 1212.493939] env[61985]: _type = "Task" [ 1212.493939] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.501750] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521e79ef-b336-3859-5b0a-c2f08c1eb402, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.557828] env[61985]: DEBUG nova.scheduler.client.report [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1212.899712] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "45cb51fd-3876-4875-9a5d-b7760072ef55" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.918389] env[61985]: DEBUG nova.network.neutron [req-7de83242-4317-4bda-8bdc-4880294bb35c req-89f5cefd-0b68-4cfd-b16e-94fdcafb6410 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updated VIF entry in instance network info cache for port 380110e4-5fc3-4dbf-b9ba-c94f348c52d2. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1212.918781] env[61985]: DEBUG nova.network.neutron [req-7de83242-4317-4bda-8bdc-4880294bb35c req-89f5cefd-0b68-4cfd-b16e-94fdcafb6410 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance_info_cache with network_info: [{"id": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "address": "fa:16:3e:56:1d:3d", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380110e4-5f", "ovs_interfaceid": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.933666] env[61985]: DEBUG oslo_concurrency.lockutils [req-39267dfa-9dc3-42c3-b056-dc19d159e8f3 req-0077cca4-8ade-4cfc-9195-36e3b4608d9e service nova] Releasing lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.004738] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]521e79ef-b336-3859-5b0a-c2f08c1eb402, 'name': SearchDatastore_Task, 'duration_secs': 0.010033} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.005092] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.005346] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1213.005602] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1213.005756] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.005953] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1213.006234] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b979d381-684d-44bf-be22-680a6d2a0bd1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.014789] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1213.014967] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1213.015686] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-044b2bad-9a03-40a9-a03f-204d8e718134 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.022839] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1213.022839] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526e3073-5b20-b62b-13e2-0e04c186b030" [ 1213.022839] env[61985]: _type = "Task" [ 1213.022839] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.030425] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526e3073-5b20-b62b-13e2-0e04c186b030, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.062341] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.194s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.422084] env[61985]: DEBUG oslo_concurrency.lockutils [req-7de83242-4317-4bda-8bdc-4880294bb35c req-89f5cefd-0b68-4cfd-b16e-94fdcafb6410 service nova] Releasing lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.429222] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d852db14-c0e1-424e-ae30-985a7334cd6f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.448389] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating instance 'f544bc65-af75-4701-802c-f6f39f39e352' progress to 0 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1213.534360] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526e3073-5b20-b62b-13e2-0e04c186b030, 'name': SearchDatastore_Task, 'duration_secs': 0.008759} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.535192] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-173e165c-c6c2-423b-a50a-6cbd17d78611 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.540974] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1213.540974] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52dd5a1f-5247-25ab-be2e-23e784823c43" [ 1213.540974] env[61985]: _type = "Task" [ 1213.540974] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.549605] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52dd5a1f-5247-25ab-be2e-23e784823c43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.570667] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c2773cbd-79e0-472b-a772-973e3df1c3ff tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.045s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.571553] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 19.269s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.571876] env[61985]: INFO nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] During sync_power_state the instance has a pending task (shelving_image_pending_upload). Skip. [ 1213.571915] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.572346] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.673s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.572528] env[61985]: INFO nova.compute.manager [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Unshelving [ 1213.954718] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1213.955108] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb5db704-c36e-4924-9f14-d5f9659c53cf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.964753] env[61985]: DEBUG oslo_vmware.api [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1213.964753] env[61985]: value = "task-936783" [ 1213.964753] env[61985]: _type = "Task" [ 1213.964753] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.973410] env[61985]: DEBUG oslo_vmware.api [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936783, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.052351] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52dd5a1f-5247-25ab-be2e-23e784823c43, 'name': SearchDatastore_Task, 'duration_secs': 0.010794} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.052632] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1214.052922] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] aa2eb44b-6c7c-41de-b9c2-0b01c2efe056/aa2eb44b-6c7c-41de-b9c2-0b01c2efe056.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1214.053207] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c7f29fa-429a-46d1-8f57-34184d9d29d3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.061088] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1214.061088] env[61985]: value = "task-936784" [ 1214.061088] env[61985]: _type = "Task" [ 1214.061088] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.070569] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.476196] env[61985]: DEBUG oslo_vmware.api [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936783, 'name': PowerOffVM_Task, 'duration_secs': 0.196852} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.476536] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1214.476810] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating instance 'f544bc65-af75-4701-802c-f6f39f39e352' progress to 17 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1214.571090] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936784, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462345} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.571350] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] aa2eb44b-6c7c-41de-b9c2-0b01c2efe056/aa2eb44b-6c7c-41de-b9c2-0b01c2efe056.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1214.571613] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1214.571890] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-444d0b83-1be8-4057-a3b8-8b0934a8504c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.577964] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1214.577964] env[61985]: value = "task-936785" [ 1214.577964] env[61985]: _type = "Task" [ 1214.577964] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.590374] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936785, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.598588] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.598869] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.599104] env[61985]: DEBUG nova.objects.instance [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'pci_requests' on Instance uuid 45cb51fd-3876-4875-9a5d-b7760072ef55 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.983853] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1214.984278] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1214.984278] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1214.984447] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1214.984563] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1214.984721] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1214.984929] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1214.985144] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1214.985441] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1214.985758] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1214.986128] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1214.994587] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09c5a38b-c192-4058-82ea-c6e540025fa2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.017193] env[61985]: DEBUG oslo_vmware.api [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1215.017193] env[61985]: value = "task-936786" [ 1215.017193] env[61985]: _type = "Task" [ 1215.017193] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.028982] env[61985]: DEBUG oslo_vmware.api [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936786, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.087932] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936785, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.19461} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.088241] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1215.089154] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37864fbc-af69-4ed6-a9f3-d09e03dbb890 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.106091] env[61985]: DEBUG nova.objects.instance [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'numa_topology' on Instance uuid 45cb51fd-3876-4875-9a5d-b7760072ef55 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.119707] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] aa2eb44b-6c7c-41de-b9c2-0b01c2efe056/aa2eb44b-6c7c-41de-b9c2-0b01c2efe056.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1215.120777] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64d7393d-7341-44ab-8836-e59356e6ce04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.143787] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1215.143787] env[61985]: value = "task-936787" [ 1215.143787] env[61985]: _type = "Task" [ 1215.143787] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.153293] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936787, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.527346] env[61985]: DEBUG oslo_vmware.api [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936786, 'name': ReconfigVM_Task, 'duration_secs': 0.305884} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.527622] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating instance 'f544bc65-af75-4701-802c-f6f39f39e352' progress to 33 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1215.621641] env[61985]: INFO nova.compute.claims [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1215.653496] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936787, 'name': ReconfigVM_Task, 'duration_secs': 0.284472} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.654414] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Reconfigured VM instance instance-00000073 to attach disk [datastore1] aa2eb44b-6c7c-41de-b9c2-0b01c2efe056/aa2eb44b-6c7c-41de-b9c2-0b01c2efe056.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1215.655053] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93a7040c-c80b-4d61-9a29-3bde5051e4aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.662123] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1215.662123] env[61985]: value = "task-936788" [ 1215.662123] env[61985]: _type = "Task" [ 1215.662123] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.670187] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936788, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.034221] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1216.034563] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1216.034743] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1216.034937] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1216.035106] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1216.035269] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1216.035483] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1216.035717] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1216.036017] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1216.036222] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1216.036417] env[61985]: DEBUG nova.virt.hardware [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1216.172687] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936788, 'name': Rename_Task, 'duration_secs': 0.133929} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.173096] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1216.173464] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5709e95-cfd0-4680-8a71-3c1cd97f7df4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.181225] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1216.181225] env[61985]: value = "task-936789" [ 1216.181225] env[61985]: _type = "Task" [ 1216.181225] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.190597] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936789, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.546370] env[61985]: ERROR nova.compute.manager [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] Traceback (most recent call last): [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] File "/opt/stack/nova/nova/compute/manager.py", line 10869, in _error_out_instance_on_exception [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] yield [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] File "/opt/stack/nova/nova/compute/manager.py", line 6111, in _resize_instance [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] disk_info = self.driver.migrate_disk_and_power_off( [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] disk_key = device.key [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] AttributeError: 'NoneType' object has no attribute 'key' [ 1216.546370] env[61985]: ERROR nova.compute.manager [instance: f544bc65-af75-4701-802c-f6f39f39e352] [ 1216.692998] env[61985]: DEBUG oslo_vmware.api [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936789, 'name': PowerOnVM_Task, 'duration_secs': 0.434229} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.693380] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1216.693597] env[61985]: INFO nova.compute.manager [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Took 6.69 seconds to spawn the instance on the hypervisor. [ 1216.693822] env[61985]: DEBUG nova.compute.manager [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1216.694615] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5997cfe4-5a61-47a8-800d-7b087b69706e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.721460] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a87b7d8-dbb0-4ce5-a628-f3fa954ad830 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.730074] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8013796-c29f-40af-9c27-141fda539959 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.762332] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361ebfef-c8af-4a60-bd0f-3d7eefa0301e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.771082] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a50d418-541f-4c2c-b92a-0e3a3f24d62d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.785109] env[61985]: DEBUG nova.compute.provider_tree [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.064260] env[61985]: INFO nova.compute.manager [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Swapping old allocation on dict_keys(['aed7e5b3-c662-4538-8447-c4f67b460215']) held by migration ef004837-0950-4c94-a03b-965efd491213 for instance [ 1217.089438] env[61985]: DEBUG nova.scheduler.client.report [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Overwriting current allocation {'allocations': {'aed7e5b3-c662-4538-8447-c4f67b460215': {'resources': {'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 163}}, 'project_id': '7bdb5f76b9b24a7ba4800a88482847da', 'user_id': '9b5abdace901430cb1e431c9a933161e', 'consumer_generation': 1} on consumer f544bc65-af75-4701-802c-f6f39f39e352 {{(pid=61985) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1217.216972] env[61985]: INFO nova.compute.manager [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Took 12.30 seconds to build instance. [ 1217.288455] env[61985]: DEBUG nova.scheduler.client.report [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1217.719603] env[61985]: DEBUG oslo_concurrency.lockutils [None req-138b953b-9239-467a-8e22-a723f8491250 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.805s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.793694] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.195s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.830132] env[61985]: INFO nova.network.neutron [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updating port eedcafdd-f3f2-4705-a528-258bb48b4574 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1218.128510] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "f544bc65-af75-4701-802c-f6f39f39e352" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.128837] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "f544bc65-af75-4701-802c-f6f39f39e352" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.129190] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "f544bc65-af75-4701-802c-f6f39f39e352-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.129504] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "f544bc65-af75-4701-802c-f6f39f39e352-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.129804] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "f544bc65-af75-4701-802c-f6f39f39e352-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.132289] env[61985]: INFO nova.compute.manager [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Terminating instance [ 1218.134241] env[61985]: DEBUG nova.compute.manager [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1218.134457] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1218.134738] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7dbd8d62-7ed4-4ce1-b14f-39f2e30f547f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.143194] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1218.143194] env[61985]: value = "task-936790" [ 1218.143194] env[61985]: _type = "Task" [ 1218.143194] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.153285] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1218.157020] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1218.157020] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211506', 'volume_id': '8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'name': 'volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f544bc65-af75-4701-802c-f6f39f39e352', 'attached_at': '', 'detached_at': '', 'volume_id': '8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'serial': '8253a6a8-8fe6-4148-adf3-55531c7a94c3'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1218.157020] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8b72b9-18f3-4eb2-9e76-f6d175aed50d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.175143] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b196a90-bcf0-454d-821c-f8b2ab0cb5fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.183402] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de0ea84-3f5e-4aae-9041-12c1a6223287 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.204343] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d48dbda-600e-41d9-8608-cb05898465ee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.227794] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] The volume has not been displaced from its original location: [datastore1] volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3/volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1218.233680] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Reconfiguring VM instance instance-00000072 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1218.234439] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-292fafbd-0ed4-4b55-8a08-d0490fccc039 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.254344] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1218.254344] env[61985]: value = "task-936791" [ 1218.254344] env[61985]: _type = "Task" [ 1218.254344] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.260506] env[61985]: DEBUG nova.compute.manager [req-fd21db6d-1cf8-472c-aac7-dc677894193e req-c9d04235-6823-4873-8ce6-32699e91c32e service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Received event network-changed-380110e4-5fc3-4dbf-b9ba-c94f348c52d2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1218.261102] env[61985]: DEBUG nova.compute.manager [req-fd21db6d-1cf8-472c-aac7-dc677894193e req-c9d04235-6823-4873-8ce6-32699e91c32e service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Refreshing instance network info cache due to event network-changed-380110e4-5fc3-4dbf-b9ba-c94f348c52d2. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1218.261300] env[61985]: DEBUG oslo_concurrency.lockutils [req-fd21db6d-1cf8-472c-aac7-dc677894193e req-c9d04235-6823-4873-8ce6-32699e91c32e service nova] Acquiring lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.262279] env[61985]: DEBUG oslo_concurrency.lockutils [req-fd21db6d-1cf8-472c-aac7-dc677894193e req-c9d04235-6823-4873-8ce6-32699e91c32e service nova] Acquired lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.262424] env[61985]: DEBUG nova.network.neutron [req-fd21db6d-1cf8-472c-aac7-dc677894193e req-c9d04235-6823-4873-8ce6-32699e91c32e service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Refreshing network info cache for port 380110e4-5fc3-4dbf-b9ba-c94f348c52d2 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1218.270156] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936791, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.643960] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.644244] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.738179] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c108003-eb06-437d-8642-4f6b0fb10493 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.746510] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfd82e1-0949-4330-ac7a-5226d0a27b8d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.783216] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905f4ff5-d78f-44c0-8bc3-0fdbd7a80d3c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.792795] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936791, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.796237] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2198e21c-59da-45a5-8305-9d27d833a9a1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.810553] env[61985]: DEBUG nova.compute.provider_tree [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1219.144364] env[61985]: DEBUG nova.network.neutron [req-fd21db6d-1cf8-472c-aac7-dc677894193e req-c9d04235-6823-4873-8ce6-32699e91c32e service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updated VIF entry in instance network info cache for port 380110e4-5fc3-4dbf-b9ba-c94f348c52d2. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1219.144753] env[61985]: DEBUG nova.network.neutron [req-fd21db6d-1cf8-472c-aac7-dc677894193e req-c9d04235-6823-4873-8ce6-32699e91c32e service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance_info_cache with network_info: [{"id": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "address": "fa:16:3e:56:1d:3d", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380110e4-5f", "ovs_interfaceid": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1219.287494] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936791, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.316022] env[61985]: DEBUG nova.scheduler.client.report [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1219.420956] env[61985]: DEBUG nova.compute.manager [req-17abae09-80f5-46bd-8286-cbdb006b4b40 req-06aa7a79-a3f0-4864-a187-afcc2965d399 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received event network-vif-plugged-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1219.421200] env[61985]: DEBUG oslo_concurrency.lockutils [req-17abae09-80f5-46bd-8286-cbdb006b4b40 req-06aa7a79-a3f0-4864-a187-afcc2965d399 service nova] Acquiring lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.421459] env[61985]: DEBUG oslo_concurrency.lockutils [req-17abae09-80f5-46bd-8286-cbdb006b4b40 req-06aa7a79-a3f0-4864-a187-afcc2965d399 service nova] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1219.421576] env[61985]: DEBUG oslo_concurrency.lockutils [req-17abae09-80f5-46bd-8286-cbdb006b4b40 req-06aa7a79-a3f0-4864-a187-afcc2965d399 service nova] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.421763] env[61985]: DEBUG nova.compute.manager [req-17abae09-80f5-46bd-8286-cbdb006b4b40 req-06aa7a79-a3f0-4864-a187-afcc2965d399 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] No waiting events found dispatching network-vif-plugged-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1219.421908] env[61985]: WARNING nova.compute.manager [req-17abae09-80f5-46bd-8286-cbdb006b4b40 req-06aa7a79-a3f0-4864-a187-afcc2965d399 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received unexpected event network-vif-plugged-eedcafdd-f3f2-4705-a528-258bb48b4574 for instance with vm_state shelved_offloaded and task_state spawning. [ 1219.505190] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1219.505410] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.505603] env[61985]: DEBUG nova.network.neutron [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1219.607570] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "f544bc65-af75-4701-802c-f6f39f39e352" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.646971] env[61985]: DEBUG oslo_concurrency.lockutils [req-fd21db6d-1cf8-472c-aac7-dc677894193e req-c9d04235-6823-4873-8ce6-32699e91c32e service nova] Releasing lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1219.788342] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936791, 'name': ReconfigVM_Task, 'duration_secs': 1.187957} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.788611] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Reconfigured VM instance instance-00000072 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1219.793216] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4daafb7a-fdaf-46ea-bc06-102318391cae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.809513] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1219.809513] env[61985]: value = "task-936792" [ 1219.809513] env[61985]: _type = "Task" [ 1219.809513] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.817989] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936792, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.819848] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.176s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.820070] env[61985]: INFO nova.compute.manager [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Successfully reverted task state from resize_migrating on failure for instance. [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server [None req-5ff5b6df-e932-419a-906c-cc22d288c5f9 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server raise self.value [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server raise self.value [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server raise self.value [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6077, in resize_instance [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server raise self.value [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6074, in resize_instance [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6111, in _resize_instance [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1219.828308] env[61985]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1219.829899] env[61985]: ERROR oslo_messaging.rpc.server [ 1220.233847] env[61985]: DEBUG nova.network.neutron [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updating instance_info_cache with network_info: [{"id": "eedcafdd-f3f2-4705-a528-258bb48b4574", "address": "fa:16:3e:49:21:b1", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeedcafdd-f3", "ovs_interfaceid": "eedcafdd-f3f2-4705-a528-258bb48b4574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.321390] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936792, 'name': ReconfigVM_Task, 'duration_secs': 0.118612} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.321695] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211506', 'volume_id': '8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'name': 'volume-8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f544bc65-af75-4701-802c-f6f39f39e352', 'attached_at': '', 'detached_at': '', 'volume_id': '8253a6a8-8fe6-4148-adf3-55531c7a94c3', 'serial': '8253a6a8-8fe6-4148-adf3-55531c7a94c3'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1220.321967] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1220.322790] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13e5e00-c48f-47cd-875f-d18006405476 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.329832] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1220.329891] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-713597e1-27bc-4dcb-bd93-79f2396869e5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.398352] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1220.398590] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1220.398814] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleting the datastore file [datastore1] f544bc65-af75-4701-802c-f6f39f39e352 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1220.399121] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3050cc0f-9a06-47a7-9854-14b33cf716e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.405809] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1220.405809] env[61985]: value = "task-936794" [ 1220.405809] env[61985]: _type = "Task" [ 1220.405809] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.413695] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936794, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.737075] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.784620] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='db71c9d4b817e87f2d356ee1a2db9e93',container_format='bare',created_at=2024-09-18T01:34:43Z,direct_url=,disk_format='vmdk',id=7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1339130780-shelved',owner='697ed07c609f4e1f86d317675a4749a8',properties=ImageMetaProps,protected=,size=31665664,status='active',tags=,updated_at=2024-09-18T01:34:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1220.784620] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1220.784839] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1220.784839] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1220.784979] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1220.785145] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1220.785368] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1220.785535] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1220.785712] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1220.785885] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1220.786127] env[61985]: DEBUG nova.virt.hardware [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1220.786971] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7edb935-a7fb-4ced-bc90-a8b35f6d505e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.795831] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e397fce5-aa5f-4820-884a-7886a7b54932 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.811987] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:21:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eedcafdd-f3f2-4705-a528-258bb48b4574', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1220.819497] env[61985]: DEBUG oslo.service.loopingcall [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1220.819782] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1220.820013] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91202a00-36b7-4772-83a2-2477f152f3ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.839625] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1220.839625] env[61985]: value = "task-936795" [ 1220.839625] env[61985]: _type = "Task" [ 1220.839625] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.847474] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936795, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.916433] env[61985]: DEBUG oslo_vmware.api [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936794, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.073544} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.916750] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1220.916968] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1220.917182] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1220.917370] env[61985]: INFO nova.compute.manager [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Took 2.78 seconds to destroy the instance on the hypervisor. [ 1220.917664] env[61985]: DEBUG oslo.service.loopingcall [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1220.917856] env[61985]: DEBUG nova.compute.manager [-] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1220.917967] env[61985]: DEBUG nova.network.neutron [-] [instance: f544bc65-af75-4701-802c-f6f39f39e352] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1220.951439] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "1b0594f2-6191-441f-b373-00b0a925d14d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1220.951712] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "1b0594f2-6191-441f-b373-00b0a925d14d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1220.952103] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "1b0594f2-6191-441f-b373-00b0a925d14d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1220.952537] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "1b0594f2-6191-441f-b373-00b0a925d14d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1220.952760] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "1b0594f2-6191-441f-b373-00b0a925d14d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.955893] env[61985]: INFO nova.compute.manager [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Terminating instance [ 1220.957943] env[61985]: DEBUG nova.compute.manager [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1220.958180] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1220.959046] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081644a3-92ab-4d35-980a-ae0a9b8799d4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.966671] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1220.966928] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e407457c-d17c-4300-b31f-92b2882f6fdf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.973442] env[61985]: DEBUG oslo_vmware.api [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1220.973442] env[61985]: value = "task-936796" [ 1220.973442] env[61985]: _type = "Task" [ 1220.973442] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.981495] env[61985]: DEBUG oslo_vmware.api [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936796, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.349396] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936795, 'name': CreateVM_Task, 'duration_secs': 0.299952} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.349769] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1221.350226] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1221.350395] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.350792] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1221.351318] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c57577aa-0728-4dca-8386-c9562365d608 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.355942] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1221.355942] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ff5ec6-c032-50ab-780d-746618172bdc" [ 1221.355942] env[61985]: _type = "Task" [ 1221.355942] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.364256] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ff5ec6-c032-50ab-780d-746618172bdc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.481451] env[61985]: DEBUG nova.compute.manager [req-3aab040e-cec4-4f7e-b26b-d1b928f51a27 req-512c1394-7d99-4e47-9cc7-03020399a56c service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received event network-changed-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1221.481594] env[61985]: DEBUG nova.compute.manager [req-3aab040e-cec4-4f7e-b26b-d1b928f51a27 req-512c1394-7d99-4e47-9cc7-03020399a56c service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Refreshing instance network info cache due to event network-changed-eedcafdd-f3f2-4705-a528-258bb48b4574. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1221.481822] env[61985]: DEBUG oslo_concurrency.lockutils [req-3aab040e-cec4-4f7e-b26b-d1b928f51a27 req-512c1394-7d99-4e47-9cc7-03020399a56c service nova] Acquiring lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1221.481973] env[61985]: DEBUG oslo_concurrency.lockutils [req-3aab040e-cec4-4f7e-b26b-d1b928f51a27 req-512c1394-7d99-4e47-9cc7-03020399a56c service nova] Acquired lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.482157] env[61985]: DEBUG nova.network.neutron [req-3aab040e-cec4-4f7e-b26b-d1b928f51a27 req-512c1394-7d99-4e47-9cc7-03020399a56c service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Refreshing network info cache for port eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1221.486581] env[61985]: DEBUG oslo_vmware.api [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936796, 'name': PowerOffVM_Task, 'duration_secs': 0.225356} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.486993] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1221.487186] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1221.487426] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d3548d5-6a54-4b81-a05e-95fe75110c6e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.560944] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1221.561207] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1221.561397] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleting the datastore file [datastore2] 1b0594f2-6191-441f-b373-00b0a925d14d {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1221.561670] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5953262f-82c1-4122-b17d-7b1466d00566 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.568020] env[61985]: DEBUG oslo_vmware.api [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1221.568020] env[61985]: value = "task-936798" [ 1221.568020] env[61985]: _type = "Task" [ 1221.568020] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.576288] env[61985]: DEBUG oslo_vmware.api [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936798, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.866173] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.866662] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Processing image 7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1221.867020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1221.867251] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.867453] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1221.867724] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d51bbb0-269e-4c1f-9bd9-db575f0a457e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.875889] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1221.876053] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1221.876766] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b969abd-e1f8-4b90-872b-6c8ff51e96cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.881682] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1221.881682] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b30b94-55ec-b225-af95-5a24d83422eb" [ 1221.881682] env[61985]: _type = "Task" [ 1221.881682] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.890095] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52b30b94-55ec-b225-af95-5a24d83422eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.009840] env[61985]: DEBUG nova.network.neutron [-] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1222.078426] env[61985]: DEBUG oslo_vmware.api [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936798, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122194} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.078494] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1222.080652] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1222.080652] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1222.080652] env[61985]: INFO nova.compute.manager [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1222.080652] env[61985]: DEBUG oslo.service.loopingcall [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1222.080652] env[61985]: DEBUG nova.compute.manager [-] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1222.080652] env[61985]: DEBUG nova.network.neutron [-] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1222.246477] env[61985]: DEBUG nova.network.neutron [req-3aab040e-cec4-4f7e-b26b-d1b928f51a27 req-512c1394-7d99-4e47-9cc7-03020399a56c service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updated VIF entry in instance network info cache for port eedcafdd-f3f2-4705-a528-258bb48b4574. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1222.246875] env[61985]: DEBUG nova.network.neutron [req-3aab040e-cec4-4f7e-b26b-d1b928f51a27 req-512c1394-7d99-4e47-9cc7-03020399a56c service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updating instance_info_cache with network_info: [{"id": "eedcafdd-f3f2-4705-a528-258bb48b4574", "address": "fa:16:3e:49:21:b1", "network": {"id": "09739b8b-d269-4c86-a7cd-b0afa9af8b0e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1522948852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "697ed07c609f4e1f86d317675a4749a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeedcafdd-f3", "ovs_interfaceid": "eedcafdd-f3f2-4705-a528-258bb48b4574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1222.392220] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Preparing fetch location {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1222.392529] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Fetch image to [datastore1] OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d/OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d.vmdk {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1222.392672] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Downloading stream optimized image 7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3 to [datastore1] OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d/OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d.vmdk on the data store datastore1 as vApp {{(pid=61985) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1222.392917] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Downloading image file data 7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3 to the ESX as VM named 'OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d' {{(pid=61985) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1222.443304] env[61985]: DEBUG nova.compute.manager [req-489ade72-196e-4bb5-a539-63caa75d6c0d req-e4b7c74b-1ab1-40bf-b35f-fdb7b0f6eb5a service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Received event network-vif-deleted-fdec9b82-21e5-4452-a82d-0c6563b85912 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1222.443510] env[61985]: INFO nova.compute.manager [req-489ade72-196e-4bb5-a539-63caa75d6c0d req-e4b7c74b-1ab1-40bf-b35f-fdb7b0f6eb5a service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Neutron deleted interface fdec9b82-21e5-4452-a82d-0c6563b85912; detaching it from the instance and deleting it from the info cache [ 1222.443684] env[61985]: DEBUG nova.network.neutron [req-489ade72-196e-4bb5-a539-63caa75d6c0d req-e4b7c74b-1ab1-40bf-b35f-fdb7b0f6eb5a service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1222.468586] env[61985]: DEBUG oslo_vmware.rw_handles [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1222.468586] env[61985]: value = "resgroup-9" [ 1222.468586] env[61985]: _type = "ResourcePool" [ 1222.468586] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1222.468887] env[61985]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-87daa14b-d556-4cb8-8692-c026e7855931 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.489827] env[61985]: DEBUG oslo_vmware.rw_handles [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lease: (returnval){ [ 1222.489827] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524541e7-20c5-2fe7-d73c-28b2767a7350" [ 1222.489827] env[61985]: _type = "HttpNfcLease" [ 1222.489827] env[61985]: } obtained for vApp import into resource pool (val){ [ 1222.489827] env[61985]: value = "resgroup-9" [ 1222.489827] env[61985]: _type = "ResourcePool" [ 1222.489827] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1222.490187] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the lease: (returnval){ [ 1222.490187] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524541e7-20c5-2fe7-d73c-28b2767a7350" [ 1222.490187] env[61985]: _type = "HttpNfcLease" [ 1222.490187] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1222.496317] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1222.496317] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524541e7-20c5-2fe7-d73c-28b2767a7350" [ 1222.496317] env[61985]: _type = "HttpNfcLease" [ 1222.496317] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1222.512181] env[61985]: INFO nova.compute.manager [-] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Took 1.59 seconds to deallocate network for instance. [ 1222.749948] env[61985]: DEBUG oslo_concurrency.lockutils [req-3aab040e-cec4-4f7e-b26b-d1b928f51a27 req-512c1394-7d99-4e47-9cc7-03020399a56c service nova] Releasing lock "refresh_cache-45cb51fd-3876-4875-9a5d-b7760072ef55" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1222.919044] env[61985]: DEBUG nova.network.neutron [-] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1222.947293] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b9f44164-0873-498e-bcdd-70c617367938 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.957029] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67824ba-3ee6-4cdb-b2d0-051e0dfc58af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.983996] env[61985]: DEBUG nova.compute.manager [req-489ade72-196e-4bb5-a539-63caa75d6c0d req-e4b7c74b-1ab1-40bf-b35f-fdb7b0f6eb5a service nova] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Detach interface failed, port_id=fdec9b82-21e5-4452-a82d-0c6563b85912, reason: Instance 1b0594f2-6191-441f-b373-00b0a925d14d could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1222.996984] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1222.996984] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524541e7-20c5-2fe7-d73c-28b2767a7350" [ 1222.996984] env[61985]: _type = "HttpNfcLease" [ 1222.996984] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1222.997300] env[61985]: DEBUG oslo_vmware.rw_handles [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1222.997300] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]524541e7-20c5-2fe7-d73c-28b2767a7350" [ 1222.997300] env[61985]: _type = "HttpNfcLease" [ 1222.997300] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1222.998014] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25922b7e-f03c-4895-83ab-c47079354408 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.006960] env[61985]: DEBUG oslo_vmware.rw_handles [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52803c1b-d192-17e8-5f44-228fd37d0337/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1223.007195] env[61985]: DEBUG oslo_vmware.rw_handles [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating HTTP connection to write to file with size = 31665664 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52803c1b-d192-17e8-5f44-228fd37d0337/disk-0.vmdk. {{(pid=61985) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1223.065717] env[61985]: INFO nova.compute.manager [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Took 0.55 seconds to detach 1 volumes for instance. [ 1223.068703] env[61985]: DEBUG nova.compute.manager [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Deleting volume: 8253a6a8-8fe6-4148-adf3-55531c7a94c3 {{(pid=61985) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3253}} [ 1223.075136] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-112ae2e1-be12-42e9-974e-eae1a778d1d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.421522] env[61985]: INFO nova.compute.manager [-] [instance: 1b0594f2-6191-441f-b373-00b0a925d14d] Took 1.34 seconds to deallocate network for instance. [ 1223.446321] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.446637] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.511907] env[61985]: DEBUG nova.compute.manager [req-cf95068f-4917-4354-a97a-7dffc6216a76 req-5fcb87e1-5874-4615-9016-a89564dc44c0 service nova] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Received event network-vif-deleted-d3381564-0201-44c0-be83-9d7de95b0fce {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1223.612085] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.612085] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.612085] env[61985]: DEBUG nova.objects.instance [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lazy-loading 'resources' on Instance uuid f544bc65-af75-4701-802c-f6f39f39e352 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1223.928573] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.950135] env[61985]: INFO nova.compute.manager [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Detaching volume 9bb2af3d-7489-4dde-b5d2-da06600326e4 [ 1223.986048] env[61985]: INFO nova.virt.block_device [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Attempting to driver detach volume 9bb2af3d-7489-4dde-b5d2-da06600326e4 from mountpoint /dev/sdb [ 1223.986549] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1223.986785] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211498', 'volume_id': '9bb2af3d-7489-4dde-b5d2-da06600326e4', 'name': 'volume-9bb2af3d-7489-4dde-b5d2-da06600326e4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '0fac6d62-46b4-49a0-ab9a-f654ef27d1f1', 'attached_at': '', 'detached_at': '', 'volume_id': '9bb2af3d-7489-4dde-b5d2-da06600326e4', 'serial': '9bb2af3d-7489-4dde-b5d2-da06600326e4'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1223.987683] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f960a8-3324-49f2-9d71-ef2dbb79250b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.016854] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fb18e0-008b-4759-adb4-f91bdab9299f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.027080] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fddaa2eb-c676-4149-b422-79f1b11b6b30 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.053732] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44f951e-1341-4fcd-8dbf-cda957cd5c4b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.069220] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] The volume has not been displaced from its original location: [datastore2] volume-9bb2af3d-7489-4dde-b5d2-da06600326e4/volume-9bb2af3d-7489-4dde-b5d2-da06600326e4.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1224.074655] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1224.076437] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b403b424-7a79-4ed8-98fa-94e738303279 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.094689] env[61985]: DEBUG oslo_vmware.api [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1224.094689] env[61985]: value = "task-936801" [ 1224.094689] env[61985]: _type = "Task" [ 1224.094689] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.107364] env[61985]: DEBUG oslo_vmware.api [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936801, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.211960] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafa0715-1937-4d16-aa38-debe983c4eba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.222350] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b5bf65-5ce8-40c0-99f8-f3f1c0b62cc7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.257358] env[61985]: DEBUG oslo_vmware.rw_handles [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Completed reading data from the image iterator. {{(pid=61985) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1224.257577] env[61985]: DEBUG oslo_vmware.rw_handles [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52803c1b-d192-17e8-5f44-228fd37d0337/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1224.258435] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6e5637-d903-4997-9e93-59f89257ae35 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.261760] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2392773f-ef63-4e16-af66-5b92458f11ae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.270614] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa41ae4-6d48-4403-b886-9679e00474d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.274388] env[61985]: DEBUG oslo_vmware.rw_handles [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52803c1b-d192-17e8-5f44-228fd37d0337/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1224.274560] env[61985]: DEBUG oslo_vmware.rw_handles [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52803c1b-d192-17e8-5f44-228fd37d0337/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1224.274819] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-951e2e2b-6df1-4e71-b79c-f76888dc385e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.286532] env[61985]: DEBUG nova.compute.provider_tree [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1224.605114] env[61985]: DEBUG oslo_vmware.api [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936801, 'name': ReconfigVM_Task, 'duration_secs': 0.243991} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.605512] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1224.610835] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9904c28a-355e-4ba5-acb8-616de3c18a0c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.620581] env[61985]: DEBUG oslo_vmware.rw_handles [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52803c1b-d192-17e8-5f44-228fd37d0337/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1224.620779] env[61985]: INFO nova.virt.vmwareapi.images [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Downloaded image file data 7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3 [ 1224.621646] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a44d4e1-5208-4a3e-b04c-0f48c80e5af0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.637044] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d85d654-4ad7-4e51-a8ce-d6cdfaa35204 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.638404] env[61985]: DEBUG oslo_vmware.api [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1224.638404] env[61985]: value = "task-936802" [ 1224.638404] env[61985]: _type = "Task" [ 1224.638404] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.647480] env[61985]: DEBUG oslo_vmware.api [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936802, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.663501] env[61985]: INFO nova.virt.vmwareapi.images [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] The imported VM was unregistered [ 1224.666135] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Caching image {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1224.666432] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Creating directory with path [datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3 {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1224.666745] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b31ace74-905a-4832-bae5-da08aa4fb800 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.676823] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Created directory with path [datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3 {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1224.677012] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d/OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d.vmdk to [datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3.vmdk. {{(pid=61985) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1224.677260] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d22e5b17-aef0-4fd6-b55b-7a73461466eb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.682558] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1224.682558] env[61985]: value = "task-936804" [ 1224.682558] env[61985]: _type = "Task" [ 1224.682558] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.689508] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936804, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.789898] env[61985]: DEBUG nova.scheduler.client.report [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1225.148293] env[61985]: DEBUG oslo_vmware.api [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936802, 'name': ReconfigVM_Task, 'duration_secs': 0.137658} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.148713] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211498', 'volume_id': '9bb2af3d-7489-4dde-b5d2-da06600326e4', 'name': 'volume-9bb2af3d-7489-4dde-b5d2-da06600326e4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '0fac6d62-46b4-49a0-ab9a-f654ef27d1f1', 'attached_at': '', 'detached_at': '', 'volume_id': '9bb2af3d-7489-4dde-b5d2-da06600326e4', 'serial': '9bb2af3d-7489-4dde-b5d2-da06600326e4'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1225.192343] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936804, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.295846] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.684s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.298348] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.370s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.298602] env[61985]: DEBUG nova.objects.instance [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lazy-loading 'resources' on Instance uuid 1b0594f2-6191-441f-b373-00b0a925d14d {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1225.316498] env[61985]: INFO nova.scheduler.client.report [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted allocations for instance f544bc65-af75-4701-802c-f6f39f39e352 [ 1225.694289] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936804, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.695788] env[61985]: DEBUG nova.objects.instance [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'flavor' on Instance uuid 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1225.826913] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4376bf45-044a-4a36-b1a6-8982dbbf6357 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "f544bc65-af75-4701-802c-f6f39f39e352" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.698s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.828485] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "f544bc65-af75-4701-802c-f6f39f39e352" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.221s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.828832] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "f544bc65-af75-4701-802c-f6f39f39e352-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.829171] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "f544bc65-af75-4701-802c-f6f39f39e352-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.829427] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "f544bc65-af75-4701-802c-f6f39f39e352-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.832058] env[61985]: INFO nova.compute.manager [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Terminating instance [ 1225.835425] env[61985]: DEBUG nova.compute.manager [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1225.836684] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-814ef423-89e0-4ba2-8a8f-a0cac7be5281 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.847739] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb6600a-ecfb-406f-9e64-e64b83e2b8bf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.883121] env[61985]: WARNING nova.virt.vmwareapi.driver [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance f544bc65-af75-4701-802c-f6f39f39e352 could not be found. [ 1225.883456] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1225.887109] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f48b2e64-bb14-427d-8c19-78ee81303de6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.898356] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34eefeb4-0b1e-47d5-be51-780b251bc735 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.933560] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f544bc65-af75-4701-802c-f6f39f39e352 could not be found. [ 1225.933789] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1225.933988] env[61985]: INFO nova.compute.manager [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Took 0.10 seconds to destroy the instance on the hypervisor. [ 1225.934288] env[61985]: DEBUG oslo.service.loopingcall [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1225.935525] env[61985]: DEBUG nova.compute.manager [-] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1225.935678] env[61985]: DEBUG nova.network.neutron [-] [instance: f544bc65-af75-4701-802c-f6f39f39e352] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1225.937879] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4725004b-375f-4533-a11a-f571b6767667 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.946246] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8efb7c-055f-485b-84f9-1c941a3118c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.979068] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc71ee18-22a8-4abb-b527-080a2998acde {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.987482] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993c853b-979b-4d28-b6b2-e828656e600a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.002233] env[61985]: DEBUG nova.compute.provider_tree [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1226.193610] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936804, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.483102] env[61985]: DEBUG nova.network.neutron [-] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.505237] env[61985]: DEBUG nova.scheduler.client.report [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1226.694523] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936804, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.703248] env[61985]: DEBUG oslo_concurrency.lockutils [None req-0e9f8031-f427-4528-aa4e-217b7fea0b88 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.256s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.985747] env[61985]: INFO nova.compute.manager [-] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Took 1.05 seconds to deallocate network for instance. [ 1226.999722] env[61985]: WARNING nova.volume.cinder [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Attachment 7d6ad0a0-e18d-47d1-9fab-7be7646ffef4 does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = 7d6ad0a0-e18d-47d1-9fab-7be7646ffef4. (HTTP 404) (Request-ID: req-00b26e8d-fe48-4ab8-ba0d-ffc7ef39b3e8) [ 1227.000034] env[61985]: INFO nova.compute.manager [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Took 0.01 seconds to detach 1 volumes for instance. [ 1227.003904] env[61985]: DEBUG nova.compute.manager [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Deleting volume: 8253a6a8-8fe6-4148-adf3-55531c7a94c3 {{(pid=61985) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3253}} [ 1227.010453] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.712s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.018143] env[61985]: WARNING nova.compute.manager [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Failed to delete volume: 8253a6a8-8fe6-4148-adf3-55531c7a94c3 due to Volume 8253a6a8-8fe6-4148-adf3-55531c7a94c3 could not be found.: nova.exception.VolumeNotFound: Volume 8253a6a8-8fe6-4148-adf3-55531c7a94c3 could not be found. [ 1227.028424] env[61985]: INFO nova.scheduler.client.report [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted allocations for instance 1b0594f2-6191-441f-b373-00b0a925d14d [ 1227.198677] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936804, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.197678} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.199561] env[61985]: INFO nova.virt.vmwareapi.ds_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d/OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d.vmdk to [datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3.vmdk. [ 1227.199762] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Cleaning up location [datastore1] OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1227.199937] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_5d16fc97-a072-422b-b6ce-1e3c27df476d {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1227.200520] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ac6717d-0571-405d-9faf-0d1f007832fe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.207299] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1227.207299] env[61985]: value = "task-936805" [ 1227.207299] env[61985]: _type = "Task" [ 1227.207299] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.215229] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936805, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.522317] env[61985]: INFO nova.compute.manager [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: f544bc65-af75-4701-802c-f6f39f39e352] Instance disappeared during terminate [ 1227.522577] env[61985]: DEBUG oslo_concurrency.lockutils [None req-6e3d3c8d-1601-4446-9c3e-32a6c60b2a3b tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "f544bc65-af75-4701-802c-f6f39f39e352" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.694s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.535236] env[61985]: DEBUG oslo_concurrency.lockutils [None req-be3921c5-2fb8-4dbf-a4e6-b3bf2e128228 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "1b0594f2-6191-441f-b373-00b0a925d14d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.583s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.689942] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.690646] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.690874] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.691079] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.691262] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.693218] env[61985]: INFO nova.compute.manager [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Terminating instance [ 1227.695010] env[61985]: DEBUG nova.compute.manager [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1227.695264] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1227.696074] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398e9421-c539-4a7b-a127-17e4d7209df3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.703488] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1227.703713] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb0aa88e-0e24-4465-bee8-b3c417be37d6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.712265] env[61985]: DEBUG oslo_vmware.api [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1227.712265] env[61985]: value = "task-936806" [ 1227.712265] env[61985]: _type = "Task" [ 1227.712265] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.718203] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936805, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033576} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.718712] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1227.718887] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.719148] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3.vmdk to [datastore1] 45cb51fd-3876-4875-9a5d-b7760072ef55/45cb51fd-3876-4875-9a5d-b7760072ef55.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1227.719378] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-061550f9-ce10-43cb-9614-27660d4a6a9e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.723608] env[61985]: DEBUG oslo_vmware.api [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936806, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.727595] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1227.727595] env[61985]: value = "task-936807" [ 1227.727595] env[61985]: _type = "Task" [ 1227.727595] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.734665] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936807, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.223416] env[61985]: DEBUG oslo_vmware.api [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936806, 'name': PowerOffVM_Task, 'duration_secs': 0.200458} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.223757] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1228.223936] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1228.224670] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a870932-26a0-42a4-bccc-fdb3caff4de2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.236996] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936807, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.291406] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1228.291629] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1228.291756] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleting the datastore file [datastore2] 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1228.292035] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c3fbcee-5fcd-4877-9416-874a36fabce6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.299265] env[61985]: DEBUG oslo_vmware.api [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1228.299265] env[61985]: value = "task-936810" [ 1228.299265] env[61985]: _type = "Task" [ 1228.299265] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.307435] env[61985]: DEBUG oslo_vmware.api [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936810, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.739071] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936807, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.811400] env[61985]: DEBUG oslo_vmware.api [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936810, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236226} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.811665] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1228.811973] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1228.812246] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1228.812445] env[61985]: INFO nova.compute.manager [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1228.812706] env[61985]: DEBUG oslo.service.loopingcall [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1228.812980] env[61985]: DEBUG nova.compute.manager [-] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1228.813103] env[61985]: DEBUG nova.network.neutron [-] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1229.240289] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936807, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.352329] env[61985]: DEBUG nova.compute.manager [req-bab2a7ac-c053-4555-853c-2e26c6cdeea4 req-bb240839-bad1-49f4-ad58-229b890a45dc service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Received event network-vif-deleted-03b6ccf4-e23e-472e-a02a-e72bd1686c90 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1229.352625] env[61985]: INFO nova.compute.manager [req-bab2a7ac-c053-4555-853c-2e26c6cdeea4 req-bb240839-bad1-49f4-ad58-229b890a45dc service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Neutron deleted interface 03b6ccf4-e23e-472e-a02a-e72bd1686c90; detaching it from the instance and deleting it from the info cache [ 1229.352824] env[61985]: DEBUG nova.network.neutron [req-bab2a7ac-c053-4555-853c-2e26c6cdeea4 req-bb240839-bad1-49f4-ad58-229b890a45dc service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.742618] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936807, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.829282] env[61985]: DEBUG nova.network.neutron [-] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.855935] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d19f1ab-bb24-4c1b-a3dc-c22caec3823a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.867049] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780b6dee-00a9-4054-b1df-ca290809bc09 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.893770] env[61985]: DEBUG nova.compute.manager [req-bab2a7ac-c053-4555-853c-2e26c6cdeea4 req-bb240839-bad1-49f4-ad58-229b890a45dc service nova] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Detach interface failed, port_id=03b6ccf4-e23e-472e-a02a-e72bd1686c90, reason: Instance 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1230.241698] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936807, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.293177} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.242747] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3/7c973ad8-9605-45ed-bd0d-9ef2c58cd5c3.vmdk to [datastore1] 45cb51fd-3876-4875-9a5d-b7760072ef55/45cb51fd-3876-4875-9a5d-b7760072ef55.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1230.242984] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5e0003-0a40-43fe-b65b-db620f45b266 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.264421] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 45cb51fd-3876-4875-9a5d-b7760072ef55/45cb51fd-3876-4875-9a5d-b7760072ef55.vmdk or device None with type streamOptimized {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1230.264683] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df593b63-042e-46e8-9d41-6ee85d34971b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.283257] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1230.283257] env[61985]: value = "task-936811" [ 1230.283257] env[61985]: _type = "Task" [ 1230.283257] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.293200] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936811, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.332435] env[61985]: INFO nova.compute.manager [-] [instance: 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1] Took 1.52 seconds to deallocate network for instance. [ 1230.657274] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "073a7668-39e6-480d-9350-835a0282b456" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.657552] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "073a7668-39e6-480d-9350-835a0282b456" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.657830] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "073a7668-39e6-480d-9350-835a0282b456-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.658061] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "073a7668-39e6-480d-9350-835a0282b456-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.658245] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "073a7668-39e6-480d-9350-835a0282b456-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.660413] env[61985]: INFO nova.compute.manager [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Terminating instance [ 1230.662254] env[61985]: DEBUG nova.compute.manager [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1230.662455] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1230.663299] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b662030-c6f3-4a5b-a495-ca37fa08c23d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.671370] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1230.671592] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-238fbf68-29d9-4a21-b94f-b5ca965eeccd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.677632] env[61985]: DEBUG oslo_vmware.api [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1230.677632] env[61985]: value = "task-936812" [ 1230.677632] env[61985]: _type = "Task" [ 1230.677632] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.685154] env[61985]: DEBUG oslo_vmware.api [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936812, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.795319] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936811, 'name': ReconfigVM_Task, 'duration_secs': 0.28543} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.795697] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 45cb51fd-3876-4875-9a5d-b7760072ef55/45cb51fd-3876-4875-9a5d-b7760072ef55.vmdk or device None with type streamOptimized {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1230.796375] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-765a22c4-95ed-4ab3-b050-8b3ffa97a6dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.802629] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1230.802629] env[61985]: value = "task-936813" [ 1230.802629] env[61985]: _type = "Task" [ 1230.802629] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.811507] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936813, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.838839] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.839223] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.839468] env[61985]: DEBUG nova.objects.instance [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'resources' on Instance uuid 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1231.188761] env[61985]: DEBUG oslo_vmware.api [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936812, 'name': PowerOffVM_Task, 'duration_secs': 0.223171} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.189116] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1231.189311] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1231.189561] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1293fa96-e96b-4736-9da1-bd2f5ac0f49f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.249122] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1231.249365] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1231.249593] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleting the datastore file [datastore2] 073a7668-39e6-480d-9350-835a0282b456 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1231.249878] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-952c857e-e12a-45fa-97c9-bb4adcc42748 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.257336] env[61985]: DEBUG oslo_vmware.api [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for the task: (returnval){ [ 1231.257336] env[61985]: value = "task-936815" [ 1231.257336] env[61985]: _type = "Task" [ 1231.257336] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.264533] env[61985]: DEBUG oslo_vmware.api [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936815, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.311303] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936813, 'name': Rename_Task, 'duration_secs': 0.153595} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.311670] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1231.311957] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45e1f4f2-63aa-4e2e-8fb7-005e21122ca9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.319107] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1231.319107] env[61985]: value = "task-936816" [ 1231.319107] env[61985]: _type = "Task" [ 1231.319107] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.324569] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936816, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.407172] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dffc0f6-b273-477a-b84c-8bdea31811f6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.414205] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca5c63c-518c-4e9c-87ce-40aae4a694c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.445108] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93441c43-8dfa-4810-a856-c7109acf73a4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.452300] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8fe56fe-e35f-450b-96c7-d9d8fc5b96c5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.465670] env[61985]: DEBUG nova.compute.provider_tree [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1231.767304] env[61985]: DEBUG oslo_vmware.api [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Task: {'id': task-936815, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173788} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.767581] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1231.767810] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1231.768023] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1231.768234] env[61985]: INFO nova.compute.manager [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] [instance: 073a7668-39e6-480d-9350-835a0282b456] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1231.768539] env[61985]: DEBUG oslo.service.loopingcall [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1231.768737] env[61985]: DEBUG nova.compute.manager [-] [instance: 073a7668-39e6-480d-9350-835a0282b456] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1231.768835] env[61985]: DEBUG nova.network.neutron [-] [instance: 073a7668-39e6-480d-9350-835a0282b456] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1231.826789] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936816, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.969629] env[61985]: DEBUG nova.scheduler.client.report [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1232.002912] env[61985]: DEBUG nova.compute.manager [req-76b0568f-d6eb-4d2d-abe6-685500206c8f req-c0d23971-4a09-4310-a1cd-6bcfd2354d86 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Received event network-vif-deleted-a5e0f361-c876-4212-8c4a-dfdfb17d5d4f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1232.003233] env[61985]: INFO nova.compute.manager [req-76b0568f-d6eb-4d2d-abe6-685500206c8f req-c0d23971-4a09-4310-a1cd-6bcfd2354d86 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Neutron deleted interface a5e0f361-c876-4212-8c4a-dfdfb17d5d4f; detaching it from the instance and deleting it from the info cache [ 1232.003324] env[61985]: DEBUG nova.network.neutron [req-76b0568f-d6eb-4d2d-abe6-685500206c8f req-c0d23971-4a09-4310-a1cd-6bcfd2354d86 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.327912] env[61985]: DEBUG oslo_vmware.api [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936816, 'name': PowerOnVM_Task, 'duration_secs': 0.654545} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.328226] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1232.425737] env[61985]: DEBUG nova.compute.manager [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1232.426670] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a945ad2a-d29c-4e06-ae91-2afba129b95d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.475778] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.636s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.481303] env[61985]: DEBUG nova.network.neutron [-] [instance: 073a7668-39e6-480d-9350-835a0282b456] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.494669] env[61985]: INFO nova.scheduler.client.report [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleted allocations for instance 0fac6d62-46b4-49a0-ab9a-f654ef27d1f1 [ 1232.506583] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3cace395-92c3-43e6-a614-edc18fcd817d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.515733] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e17e0cc-9f13-4fc0-8b12-dd5a57cc2cde {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.541516] env[61985]: DEBUG nova.compute.manager [req-76b0568f-d6eb-4d2d-abe6-685500206c8f req-c0d23971-4a09-4310-a1cd-6bcfd2354d86 service nova] [instance: 073a7668-39e6-480d-9350-835a0282b456] Detach interface failed, port_id=a5e0f361-c876-4212-8c4a-dfdfb17d5d4f, reason: Instance 073a7668-39e6-480d-9350-835a0282b456 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1232.943909] env[61985]: DEBUG oslo_concurrency.lockutils [None req-3f1d8695-9d80-4e6a-ab0f-8f0d4fad6887 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.371s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.983773] env[61985]: INFO nova.compute.manager [-] [instance: 073a7668-39e6-480d-9350-835a0282b456] Took 1.21 seconds to deallocate network for instance. [ 1233.000790] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a0a9c7f8-8eda-491e-b818-d7b03bca6efb tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "0fac6d62-46b4-49a0-ab9a-f654ef27d1f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.310s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1233.491518] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.491518] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1233.491518] env[61985]: DEBUG nova.objects.instance [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lazy-loading 'resources' on Instance uuid 073a7668-39e6-480d-9350-835a0282b456 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1234.047552] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bd6f46-3359-43da-a0c6-93114f8dfe5a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.055384] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e950a72-9079-4759-b319-b96f40471767 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.085049] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1fe84c-9a83-473e-8a10-d1581e9a5443 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.092466] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49065261-cbcb-46d3-83ac-1b7a1210653b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.107318] env[61985]: DEBUG nova.compute.provider_tree [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1234.369314] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "45cb51fd-3876-4875-9a5d-b7760072ef55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.369591] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.369802] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.369991] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.370188] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.372253] env[61985]: INFO nova.compute.manager [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Terminating instance [ 1234.374377] env[61985]: DEBUG nova.compute.manager [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1234.374580] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1234.375451] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4b3095-6d38-4443-8f09-24fede5396dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.382929] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1234.383179] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6589c644-7bb9-4a3f-92e7-2fed89262fa1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.389765] env[61985]: DEBUG oslo_vmware.api [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1234.389765] env[61985]: value = "task-936817" [ 1234.389765] env[61985]: _type = "Task" [ 1234.389765] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.397154] env[61985]: DEBUG oslo_vmware.api [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936817, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.610614] env[61985]: DEBUG nova.scheduler.client.report [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1234.827863] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "8c95afa2-dbec-4408-a27d-595431e09748" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.828155] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "8c95afa2-dbec-4408-a27d-595431e09748" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.899582] env[61985]: DEBUG oslo_vmware.api [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936817, 'name': PowerOffVM_Task, 'duration_secs': 0.210038} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.899867] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1234.900130] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1234.900388] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-515d0f77-2ce8-44d6-9938-ec8e94636c25 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.030743] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1235.030984] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1235.031192] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleting the datastore file [datastore1] 45cb51fd-3876-4875-9a5d-b7760072ef55 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1235.031459] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4bfb5ee-9305-4840-b131-fadfb33a610a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.037798] env[61985]: DEBUG oslo_vmware.api [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for the task: (returnval){ [ 1235.037798] env[61985]: value = "task-936819" [ 1235.037798] env[61985]: _type = "Task" [ 1235.037798] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.045605] env[61985]: DEBUG oslo_vmware.api [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936819, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.115621] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.625s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.138640] env[61985]: INFO nova.scheduler.client.report [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Deleted allocations for instance 073a7668-39e6-480d-9350-835a0282b456 [ 1235.331126] env[61985]: DEBUG nova.compute.manager [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1235.547465] env[61985]: DEBUG oslo_vmware.api [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Task: {'id': task-936819, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126688} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.547742] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1235.547915] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1235.548109] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1235.548290] env[61985]: INFO nova.compute.manager [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1235.548534] env[61985]: DEBUG oslo.service.loopingcall [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1235.548751] env[61985]: DEBUG nova.compute.manager [-] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1235.548844] env[61985]: DEBUG nova.network.neutron [-] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1235.647381] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a125f092-869d-4bed-b839-f1520340dcd2 tempest-ServerActionsTestOtherA-2137190543 tempest-ServerActionsTestOtherA-2137190543-project-member] Lock "073a7668-39e6-480d-9350-835a0282b456" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.989s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.857408] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.857711] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.859238] env[61985]: INFO nova.compute.claims [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1235.970421] env[61985]: DEBUG nova.compute.manager [req-dd693307-99d1-4f20-87df-bb0ba4f32c69 req-a03720d7-1270-4565-a71d-da24025ee7d7 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Received event network-vif-deleted-eedcafdd-f3f2-4705-a528-258bb48b4574 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1235.970546] env[61985]: INFO nova.compute.manager [req-dd693307-99d1-4f20-87df-bb0ba4f32c69 req-a03720d7-1270-4565-a71d-da24025ee7d7 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Neutron deleted interface eedcafdd-f3f2-4705-a528-258bb48b4574; detaching it from the instance and deleting it from the info cache [ 1235.970729] env[61985]: DEBUG nova.network.neutron [req-dd693307-99d1-4f20-87df-bb0ba4f32c69 req-a03720d7-1270-4565-a71d-da24025ee7d7 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1236.448822] env[61985]: DEBUG nova.network.neutron [-] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1236.475558] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4175790c-de9f-48f7-bf00-36d58cc90bb6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.484303] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8041ee41-039f-4ad4-9ac8-24acf31663f3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.507983] env[61985]: DEBUG nova.compute.manager [req-dd693307-99d1-4f20-87df-bb0ba4f32c69 req-a03720d7-1270-4565-a71d-da24025ee7d7 service nova] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Detach interface failed, port_id=eedcafdd-f3f2-4705-a528-258bb48b4574, reason: Instance 45cb51fd-3876-4875-9a5d-b7760072ef55 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1236.919593] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed810cec-1ebb-4dba-af83-396bd862452c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.927288] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14ff514-af4c-432d-9909-dd35eeb77080 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.955819] env[61985]: INFO nova.compute.manager [-] [instance: 45cb51fd-3876-4875-9a5d-b7760072ef55] Took 1.41 seconds to deallocate network for instance. [ 1236.958153] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627d719c-af71-4698-8a37-6cada2a0a640 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.967883] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334c4c02-2e2f-41e0-b6b9-bc540dee5b76 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.983212] env[61985]: DEBUG nova.compute.provider_tree [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1237.465259] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1237.486771] env[61985]: DEBUG nova.scheduler.client.report [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1237.991954] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.134s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1237.992770] env[61985]: DEBUG nova.compute.manager [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1237.996061] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.531s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.996308] env[61985]: DEBUG nova.objects.instance [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lazy-loading 'resources' on Instance uuid 45cb51fd-3876-4875-9a5d-b7760072ef55 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1238.499703] env[61985]: DEBUG nova.compute.utils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1238.504311] env[61985]: DEBUG nova.compute.manager [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1238.504425] env[61985]: DEBUG nova.network.neutron [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1238.552897] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2515b8db-e3ee-4534-b8ac-05a67b0edc77 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.556841] env[61985]: DEBUG nova.policy [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b489c06ceada46bea5190b44dafd777a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e7e22f7d7b7b407ebb2d00e174d8a42c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1238.563132] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d418c46e-9eab-4aea-8184-9c5dfcda02bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.592727] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b111a336-df05-494f-8b45-65523c59662e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.599868] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7017a04d-80c1-4548-b20f-5c97d74344b5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.614219] env[61985]: DEBUG nova.compute.provider_tree [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1238.833104] env[61985]: DEBUG nova.network.neutron [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Successfully created port: a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1239.006423] env[61985]: DEBUG nova.compute.manager [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1239.117796] env[61985]: DEBUG nova.scheduler.client.report [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1239.622990] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.627s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.645832] env[61985]: INFO nova.scheduler.client.report [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Deleted allocations for instance 45cb51fd-3876-4875-9a5d-b7760072ef55 [ 1240.015726] env[61985]: DEBUG nova.compute.manager [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1240.040980] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1240.041373] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1240.041552] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1240.041747] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1240.041903] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1240.042074] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1240.042296] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1240.042464] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1240.042639] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1240.042809] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1240.042990] env[61985]: DEBUG nova.virt.hardware [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1240.043863] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1dbad5-86f0-4e8f-8a2b-22c5ed8126fb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.051890] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cbc0dce-7171-47bc-978a-894c1101f8c1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.152901] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17dc5088-6688-457d-8215-a955d005b494 tempest-ServerActionsTestOtherB-235429838 tempest-ServerActionsTestOtherB-235429838-project-member] Lock "45cb51fd-3876-4875-9a5d-b7760072ef55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.783s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.233964] env[61985]: DEBUG nova.compute.manager [req-a038756b-2b8a-453e-b0c0-074a8e37831e req-1c8f4887-cd53-41ab-a7d1-07148579ad75 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received event network-vif-plugged-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1240.234221] env[61985]: DEBUG oslo_concurrency.lockutils [req-a038756b-2b8a-453e-b0c0-074a8e37831e req-1c8f4887-cd53-41ab-a7d1-07148579ad75 service nova] Acquiring lock "8c95afa2-dbec-4408-a27d-595431e09748-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.234445] env[61985]: DEBUG oslo_concurrency.lockutils [req-a038756b-2b8a-453e-b0c0-074a8e37831e req-1c8f4887-cd53-41ab-a7d1-07148579ad75 service nova] Lock "8c95afa2-dbec-4408-a27d-595431e09748-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.234622] env[61985]: DEBUG oslo_concurrency.lockutils [req-a038756b-2b8a-453e-b0c0-074a8e37831e req-1c8f4887-cd53-41ab-a7d1-07148579ad75 service nova] Lock "8c95afa2-dbec-4408-a27d-595431e09748-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.234800] env[61985]: DEBUG nova.compute.manager [req-a038756b-2b8a-453e-b0c0-074a8e37831e req-1c8f4887-cd53-41ab-a7d1-07148579ad75 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] No waiting events found dispatching network-vif-plugged-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1240.234974] env[61985]: WARNING nova.compute.manager [req-a038756b-2b8a-453e-b0c0-074a8e37831e req-1c8f4887-cd53-41ab-a7d1-07148579ad75 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received unexpected event network-vif-plugged-a71d390e-e1df-4d20-a37a-e5a2331f9b7f for instance with vm_state building and task_state spawning. [ 1240.779707] env[61985]: DEBUG nova.network.neutron [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Successfully updated port: a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1240.804113] env[61985]: DEBUG nova.compute.manager [req-5ff761d0-cb56-46eb-87dc-3138c40d02e3 req-b42fe2c4-b496-4956-86ab-98e5a63952ac service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received event network-changed-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1240.804425] env[61985]: DEBUG nova.compute.manager [req-5ff761d0-cb56-46eb-87dc-3138c40d02e3 req-b42fe2c4-b496-4956-86ab-98e5a63952ac service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Refreshing instance network info cache due to event network-changed-a71d390e-e1df-4d20-a37a-e5a2331f9b7f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1240.804733] env[61985]: DEBUG oslo_concurrency.lockutils [req-5ff761d0-cb56-46eb-87dc-3138c40d02e3 req-b42fe2c4-b496-4956-86ab-98e5a63952ac service nova] Acquiring lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1240.804921] env[61985]: DEBUG oslo_concurrency.lockutils [req-5ff761d0-cb56-46eb-87dc-3138c40d02e3 req-b42fe2c4-b496-4956-86ab-98e5a63952ac service nova] Acquired lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.805143] env[61985]: DEBUG nova.network.neutron [req-5ff761d0-cb56-46eb-87dc-3138c40d02e3 req-b42fe2c4-b496-4956-86ab-98e5a63952ac service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Refreshing network info cache for port a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1241.282773] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1241.338608] env[61985]: DEBUG nova.network.neutron [req-5ff761d0-cb56-46eb-87dc-3138c40d02e3 req-b42fe2c4-b496-4956-86ab-98e5a63952ac service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1241.414454] env[61985]: DEBUG nova.network.neutron [req-5ff761d0-cb56-46eb-87dc-3138c40d02e3 req-b42fe2c4-b496-4956-86ab-98e5a63952ac service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1241.917615] env[61985]: DEBUG oslo_concurrency.lockutils [req-5ff761d0-cb56-46eb-87dc-3138c40d02e3 req-b42fe2c4-b496-4956-86ab-98e5a63952ac service nova] Releasing lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1241.918054] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.918225] env[61985]: DEBUG nova.network.neutron [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1242.449217] env[61985]: DEBUG nova.network.neutron [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1242.573727] env[61985]: DEBUG nova.network.neutron [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating instance_info_cache with network_info: [{"id": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "address": "fa:16:3e:e1:8c:42", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71d390e-e1", "ovs_interfaceid": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1243.076587] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1243.076927] env[61985]: DEBUG nova.compute.manager [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Instance network_info: |[{"id": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "address": "fa:16:3e:e1:8c:42", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71d390e-e1", "ovs_interfaceid": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1243.077393] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:8c:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e55c248-c504-4c7a-bbe9-f42cf417aee7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a71d390e-e1df-4d20-a37a-e5a2331f9b7f', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1243.085295] env[61985]: DEBUG oslo.service.loopingcall [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1243.085505] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1243.085765] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66c39368-1836-4f55-8706-7e60ae7f2dbe {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.106279] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1243.106279] env[61985]: value = "task-936821" [ 1243.106279] env[61985]: _type = "Task" [ 1243.106279] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.113684] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936821, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.616374] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936821, 'name': CreateVM_Task, 'duration_secs': 0.301531} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.616574] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1243.617219] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1243.617388] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.617727] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1243.617980] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d82541cb-110f-42be-8514-3203976b51ca {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.622362] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1243.622362] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e189ec-353e-8f60-c6c1-f2dd9bbe8884" [ 1243.622362] env[61985]: _type = "Task" [ 1243.622362] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.629553] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e189ec-353e-8f60-c6c1-f2dd9bbe8884, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.133106] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e189ec-353e-8f60-c6c1-f2dd9bbe8884, 'name': SearchDatastore_Task, 'duration_secs': 0.009774} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.133756] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1244.134130] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1244.134479] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.134825] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.135138] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1244.135499] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b01d6df4-30fd-495b-b252-60f01f1b5302 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.144028] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1244.144028] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1244.144503] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8fa5f5a-830a-4162-b42d-cb511f17db85 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.149569] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1244.149569] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d8f906-a659-9359-9330-88b04b63bb44" [ 1244.149569] env[61985]: _type = "Task" [ 1244.149569] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.157925] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d8f906-a659-9359-9330-88b04b63bb44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.659547] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52d8f906-a659-9359-9330-88b04b63bb44, 'name': SearchDatastore_Task, 'duration_secs': 0.008637} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.660291] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0977b7a6-be7e-4717-98ef-95225a47d837 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.665138] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1244.665138] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522cc309-a199-c650-208a-b9658c2cdfb7" [ 1244.665138] env[61985]: _type = "Task" [ 1244.665138] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.672407] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522cc309-a199-c650-208a-b9658c2cdfb7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.177461] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]522cc309-a199-c650-208a-b9658c2cdfb7, 'name': SearchDatastore_Task, 'duration_secs': 0.009275} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.177804] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.178158] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8c95afa2-dbec-4408-a27d-595431e09748/8c95afa2-dbec-4408-a27d-595431e09748.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1245.178437] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4a12c9f-86f2-4fb9-8e2e-cccaad1b56f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.185384] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1245.185384] env[61985]: value = "task-936822" [ 1245.185384] env[61985]: _type = "Task" [ 1245.185384] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.193371] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936822, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.695121] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936822, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.197374] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936822, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511118} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.197690] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 8c95afa2-dbec-4408-a27d-595431e09748/8c95afa2-dbec-4408-a27d-595431e09748.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1246.197839] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1246.198124] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-524ca887-e52c-4cd1-bd2f-628c68df10ed {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.204701] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1246.204701] env[61985]: value = "task-936823" [ 1246.204701] env[61985]: _type = "Task" [ 1246.204701] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.213510] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936823, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.714857] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936823, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068592} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.715141] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1246.715913] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246461ec-742b-4822-838e-6f862ffec938 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.737306] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Reconfiguring VM instance instance-00000074 to attach disk [datastore2] 8c95afa2-dbec-4408-a27d-595431e09748/8c95afa2-dbec-4408-a27d-595431e09748.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1246.737504] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37256392-f963-4edd-9103-610963ee781f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.755807] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1246.755807] env[61985]: value = "task-936824" [ 1246.755807] env[61985]: _type = "Task" [ 1246.755807] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.763406] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936824, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.268201] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936824, 'name': ReconfigVM_Task, 'duration_secs': 0.268763} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.268201] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Reconfigured VM instance instance-00000074 to attach disk [datastore2] 8c95afa2-dbec-4408-a27d-595431e09748/8c95afa2-dbec-4408-a27d-595431e09748.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1247.268201] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-899912b1-c212-4428-869c-0f781164e8d1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.273017] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1247.273017] env[61985]: value = "task-936825" [ 1247.273017] env[61985]: _type = "Task" [ 1247.273017] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.281801] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936825, 'name': Rename_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.783370] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936825, 'name': Rename_Task} progress is 14%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.284550] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936825, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.785380] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936825, 'name': Rename_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.288574] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936825, 'name': Rename_Task, 'duration_secs': 1.831595} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.288574] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1249.288574] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d5a0415-e3b1-46af-b752-b3511e6fa71f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.295091] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1249.295091] env[61985]: value = "task-936829" [ 1249.295091] env[61985]: _type = "Task" [ 1249.295091] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.300495] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936829, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.807246] env[61985]: DEBUG oslo_vmware.api [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936829, 'name': PowerOnVM_Task, 'duration_secs': 0.490152} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.807625] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1249.807837] env[61985]: INFO nova.compute.manager [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Took 9.79 seconds to spawn the instance on the hypervisor. [ 1249.808154] env[61985]: DEBUG nova.compute.manager [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1249.809074] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967fbdff-f966-4331-bd06-728f54a6e65c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.330033] env[61985]: INFO nova.compute.manager [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Took 14.49 seconds to build instance. [ 1250.832469] env[61985]: DEBUG oslo_concurrency.lockutils [None req-776ffd21-3b3c-42c3-aa37-1864b798ff50 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "8c95afa2-dbec-4408-a27d-595431e09748" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.004s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1251.562374] env[61985]: DEBUG nova.compute.manager [req-a1ef917b-7dab-4a01-a443-34095c22ee86 req-515d4f2f-6e6c-474b-aa4f-bf3be3e69fb9 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received event network-changed-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1251.562674] env[61985]: DEBUG nova.compute.manager [req-a1ef917b-7dab-4a01-a443-34095c22ee86 req-515d4f2f-6e6c-474b-aa4f-bf3be3e69fb9 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Refreshing instance network info cache due to event network-changed-a71d390e-e1df-4d20-a37a-e5a2331f9b7f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1251.563687] env[61985]: DEBUG oslo_concurrency.lockutils [req-a1ef917b-7dab-4a01-a443-34095c22ee86 req-515d4f2f-6e6c-474b-aa4f-bf3be3e69fb9 service nova] Acquiring lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1251.563687] env[61985]: DEBUG oslo_concurrency.lockutils [req-a1ef917b-7dab-4a01-a443-34095c22ee86 req-515d4f2f-6e6c-474b-aa4f-bf3be3e69fb9 service nova] Acquired lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.563687] env[61985]: DEBUG nova.network.neutron [req-a1ef917b-7dab-4a01-a443-34095c22ee86 req-515d4f2f-6e6c-474b-aa4f-bf3be3e69fb9 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Refreshing network info cache for port a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1252.288307] env[61985]: DEBUG nova.network.neutron [req-a1ef917b-7dab-4a01-a443-34095c22ee86 req-515d4f2f-6e6c-474b-aa4f-bf3be3e69fb9 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updated VIF entry in instance network info cache for port a71d390e-e1df-4d20-a37a-e5a2331f9b7f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1252.288745] env[61985]: DEBUG nova.network.neutron [req-a1ef917b-7dab-4a01-a443-34095c22ee86 req-515d4f2f-6e6c-474b-aa4f-bf3be3e69fb9 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating instance_info_cache with network_info: [{"id": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "address": "fa:16:3e:e1:8c:42", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71d390e-e1", "ovs_interfaceid": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1252.791610] env[61985]: DEBUG oslo_concurrency.lockutils [req-a1ef917b-7dab-4a01-a443-34095c22ee86 req-515d4f2f-6e6c-474b-aa4f-bf3be3e69fb9 service nova] Releasing lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.499661] env[61985]: DEBUG nova.compute.manager [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Stashing vm_state: active {{(pid=61985) _prep_resize /opt/stack/nova/nova/compute/manager.py:5630}} [ 1257.019651] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1257.019937] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1257.525114] env[61985]: INFO nova.compute.claims [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1257.898543] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1257.898787] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1257.898879] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1258.031487] env[61985]: INFO nova.compute.resource_tracker [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating resource usage from migration b3baf538-5899-47cc-9e45-8f0fe1e077ee [ 1258.077966] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237bc843-df80-40bf-848f-65741f6fb004 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.085370] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55db7a0-b4f2-45e3-861f-15d49850bbb8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.117181] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e54c03-ff13-4b00-b21c-29fbf8ed3a2f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.125128] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65501a86-eee4-4117-9e31-cde53f966798 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.138058] env[61985]: DEBUG nova.compute.provider_tree [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1258.643761] env[61985]: DEBUG nova.scheduler.client.report [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1259.149954] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.129s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1259.149954] env[61985]: INFO nova.compute.manager [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Migrating [ 1259.664595] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1259.664971] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1259.664971] env[61985]: DEBUG nova.network.neutron [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1260.381274] env[61985]: DEBUG nova.network.neutron [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance_info_cache with network_info: [{"id": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "address": "fa:16:3e:56:1d:3d", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380110e4-5f", "ovs_interfaceid": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1260.884298] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1260.915506] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Didn't find any instances for network info cache update. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10018}} [ 1260.915725] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1260.915873] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1260.916033] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1260.916190] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1260.916482] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1260.916482] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1260.916703] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 1260.916766] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.419627] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.419880] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1261.420071] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.420230] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1261.421151] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646a1719-027f-4a07-b6d6-efe579c99f38 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.429673] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ece6408-b4e6-4f72-8056-f065cd75b00f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.443603] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04d0823-61fa-4f41-a1fe-638f6f974e20 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.449607] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b995e4f9-e7ca-4e9c-9f0b-481a19fd50c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.477888] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181172MB free_disk=43GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1261.478036] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.478227] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.402046] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0316295b-1d07-41c1-b9c5-d6346d1ac99c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.421928] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance 'aa2eb44b-6c7c-41de-b9c2-0b01c2efe056' progress to 0 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1262.484063] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Applying migration context for instance aa2eb44b-6c7c-41de-b9c2-0b01c2efe056 as it has an incoming, in-progress migration b3baf538-5899-47cc-9e45-8f0fe1e077ee. Migration status is migrating {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1262.484730] env[61985]: INFO nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating resource usage from migration b3baf538-5899-47cc-9e45-8f0fe1e077ee [ 1262.499057] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8c95afa2-dbec-4408-a27d-595431e09748 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1262.499223] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Migration b3baf538-5899-47cc-9e45-8f0fe1e077ee is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1262.499328] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance aa2eb44b-6c7c-41de-b9c2-0b01c2efe056 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1262.499501] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1262.499637] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1152MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1262.545030] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968aab96-063d-458d-8c88-4e7dce861be2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.553097] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23eb3d28-bd1c-4e0d-b5e2-2180a6357288 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.582545] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f5508c-80a4-4f1e-bfbf-65627799006b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.589708] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37eb9991-bff0-4edf-ae03-dc2133cb46ec {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.603658] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1262.928012] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1262.928371] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-640397db-c3d2-4328-ba8b-7cdcd3c539ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.936901] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1262.936901] env[61985]: value = "task-936836" [ 1262.936901] env[61985]: _type = "Task" [ 1262.936901] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.948933] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936836, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.106408] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1263.447078] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936836, 'name': PowerOffVM_Task, 'duration_secs': 0.231667} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.447528] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1263.447528] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance 'aa2eb44b-6c7c-41de-b9c2-0b01c2efe056' progress to 17 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1263.611013] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1263.611196] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.133s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1263.954438] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1263.954684] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1263.954842] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1263.955048] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1263.955213] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1263.955373] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1263.955584] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1263.955758] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1263.955929] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1263.956114] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1263.956298] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1263.961327] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f12de574-3be9-4b7c-9a35-5431a081f458 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.978762] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1263.978762] env[61985]: value = "task-936837" [ 1263.978762] env[61985]: _type = "Task" [ 1263.978762] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.986820] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936837, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.488959] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936837, 'name': ReconfigVM_Task, 'duration_secs': 0.168857} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.489388] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance 'aa2eb44b-6c7c-41de-b9c2-0b01c2efe056' progress to 33 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1264.996136] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1264.996361] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1264.996524] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1264.996754] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1264.996941] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1264.997111] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1264.997326] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1264.997491] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1264.997660] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1264.997863] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1264.998056] env[61985]: DEBUG nova.virt.hardware [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1265.003466] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Reconfiguring VM instance instance-00000073 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1265.003751] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91c37546-d4f1-48bb-92bc-af3eff264c22 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.023406] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1265.023406] env[61985]: value = "task-936839" [ 1265.023406] env[61985]: _type = "Task" [ 1265.023406] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.031819] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936839, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.533209] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936839, 'name': ReconfigVM_Task, 'duration_secs': 0.161699} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.533535] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Reconfigured VM instance instance-00000073 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1265.534225] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-083bcf6b-1652-4331-8ff2-bd0b65181874 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.555411] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] aa2eb44b-6c7c-41de-b9c2-0b01c2efe056/aa2eb44b-6c7c-41de-b9c2-0b01c2efe056.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1265.555960] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a7fc453-7f9e-42a5-8e0e-bb700a4caed7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.575560] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1265.575560] env[61985]: value = "task-936840" [ 1265.575560] env[61985]: _type = "Task" [ 1265.575560] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.584809] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936840, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.085967] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936840, 'name': ReconfigVM_Task, 'duration_secs': 0.25378} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.086266] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Reconfigured VM instance instance-00000073 to attach disk [datastore1] aa2eb44b-6c7c-41de-b9c2-0b01c2efe056/aa2eb44b-6c7c-41de-b9c2-0b01c2efe056.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1266.086525] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance 'aa2eb44b-6c7c-41de-b9c2-0b01c2efe056' progress to 50 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1266.592970] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fbb620-33fc-4446-bdf4-dd4530223857 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.613045] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3288a0b8-f8a6-4227-83ef-19e5b4fd3b1a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.630091] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance 'aa2eb44b-6c7c-41de-b9c2-0b01c2efe056' progress to 67 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1267.169241] env[61985]: DEBUG nova.network.neutron [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Port 380110e4-5fc3-4dbf-b9ba-c94f348c52d2 binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1268.192163] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.192163] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.192803] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.499646] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Acquiring lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.499901] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.002570] env[61985]: DEBUG nova.compute.manager [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1269.228030] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1269.228343] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1269.228400] env[61985]: DEBUG nova.network.neutron [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1269.523116] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.523405] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.525008] env[61985]: INFO nova.compute.claims [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1269.988046] env[61985]: DEBUG nova.network.neutron [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance_info_cache with network_info: [{"id": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "address": "fa:16:3e:56:1d:3d", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380110e4-5f", "ovs_interfaceid": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1270.490756] env[61985]: DEBUG oslo_concurrency.lockutils [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1270.589663] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f73a753-a13d-4394-af44-0e51ffc6c3af {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.597984] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116bb4f4-2435-485c-884a-a2955764fa04 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.628276] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56af54e2-ec9b-4d3c-82e3-6e54e69c86c4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.635144] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb187c4f-0ab1-4d6d-a107-8375324e3a07 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.328980] env[61985]: DEBUG nova.compute.provider_tree [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1271.333854] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93a2434-9fc2-48f4-a47d-f65bf280facb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.350779] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3815e91c-8b5a-496f-b0f1-f2d889ae1363 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.358435] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance 'aa2eb44b-6c7c-41de-b9c2-0b01c2efe056' progress to 83 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1271.835608] env[61985]: DEBUG nova.scheduler.client.report [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1271.864257] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1271.864544] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d7a2764-58b2-4498-9892-5938b62dbb4b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.872811] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1271.872811] env[61985]: value = "task-936841" [ 1271.872811] env[61985]: _type = "Task" [ 1271.872811] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.880962] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936841, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.340593] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.817s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1272.341167] env[61985]: DEBUG nova.compute.manager [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1272.382963] env[61985]: DEBUG oslo_vmware.api [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936841, 'name': PowerOnVM_Task, 'duration_secs': 0.361106} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.383219] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1272.383407] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-1590ebc3-771f-4d12-93cc-72a7a151689f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance 'aa2eb44b-6c7c-41de-b9c2-0b01c2efe056' progress to 100 {{(pid=61985) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1272.847212] env[61985]: DEBUG nova.compute.utils [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1272.848757] env[61985]: DEBUG nova.compute.manager [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1272.848954] env[61985]: DEBUG nova.network.neutron [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1272.886850] env[61985]: DEBUG nova.policy [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '87817ee92a954b1392c81ace74cd85e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e55fe9a13574d95828bbf020d7ee62e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1273.167742] env[61985]: DEBUG nova.network.neutron [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Successfully created port: 8138f233-84b3-454c-a61a-4de873fa461b {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1273.352787] env[61985]: DEBUG nova.compute.manager [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1273.858684] env[61985]: INFO nova.virt.block_device [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Booting with volume a1fed62e-6782-4ff4-b5de-21f1669862fc at /dev/sda [ 1273.893597] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18ddb135-00ee-4c4c-b46e-728ca4cfcc57 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.910373] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64bc781-0f4d-4064-a2de-f2e8588f6432 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.938019] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e39ef73-2946-455c-879f-7734cf0a290f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.946523] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882a0d15-314e-4320-8315-72d21796a943 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.969753] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7894f0fb-01f1-45ac-9348-f52a68c7fecb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.976218] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbfb8231-664f-413d-987f-c087d150da4d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.989922] env[61985]: DEBUG nova.virt.block_device [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Updating existing volume attachment record: 88eab190-09d2-431d-b7e7-a9742ab07235 {{(pid=61985) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1274.734180] env[61985]: DEBUG nova.compute.manager [req-21372b3c-6eaa-457a-a9b4-2e29d822f4df req-9a8ae653-65c9-4583-813b-57894ed36b54 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Received event network-vif-plugged-8138f233-84b3-454c-a61a-4de873fa461b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1274.734180] env[61985]: DEBUG oslo_concurrency.lockutils [req-21372b3c-6eaa-457a-a9b4-2e29d822f4df req-9a8ae653-65c9-4583-813b-57894ed36b54 service nova] Acquiring lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1274.735375] env[61985]: DEBUG oslo_concurrency.lockutils [req-21372b3c-6eaa-457a-a9b4-2e29d822f4df req-9a8ae653-65c9-4583-813b-57894ed36b54 service nova] Lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1274.736436] env[61985]: DEBUG oslo_concurrency.lockutils [req-21372b3c-6eaa-457a-a9b4-2e29d822f4df req-9a8ae653-65c9-4583-813b-57894ed36b54 service nova] Lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1274.736838] env[61985]: DEBUG nova.compute.manager [req-21372b3c-6eaa-457a-a9b4-2e29d822f4df req-9a8ae653-65c9-4583-813b-57894ed36b54 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] No waiting events found dispatching network-vif-plugged-8138f233-84b3-454c-a61a-4de873fa461b {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1274.737225] env[61985]: WARNING nova.compute.manager [req-21372b3c-6eaa-457a-a9b4-2e29d822f4df req-9a8ae653-65c9-4583-813b-57894ed36b54 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Received unexpected event network-vif-plugged-8138f233-84b3-454c-a61a-4de873fa461b for instance with vm_state building and task_state block_device_mapping. [ 1274.881897] env[61985]: DEBUG nova.network.neutron [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Successfully updated port: 8138f233-84b3-454c-a61a-4de873fa461b {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1275.316142] env[61985]: DEBUG nova.network.neutron [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Port 380110e4-5fc3-4dbf-b9ba-c94f348c52d2 binding to destination host cpu-1 is already ACTIVE {{(pid=61985) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1275.316408] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1275.316565] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1275.316732] env[61985]: DEBUG nova.network.neutron [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1275.384574] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Acquiring lock "refresh_cache-d74cbed6-8f70-4e8d-b050-23e00ac334d4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1275.384770] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Acquired lock "refresh_cache-d74cbed6-8f70-4e8d-b050-23e00ac334d4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1275.384875] env[61985]: DEBUG nova.network.neutron [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1275.917479] env[61985]: DEBUG nova.network.neutron [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1276.046707] env[61985]: DEBUG nova.network.neutron [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Updating instance_info_cache with network_info: [{"id": "8138f233-84b3-454c-a61a-4de873fa461b", "address": "fa:16:3e:b4:0f:de", "network": {"id": "9240338e-ddc4-4cce-b5b5-bb259be32858", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-565887658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e55fe9a13574d95828bbf020d7ee62e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8138f233-84", "ovs_interfaceid": "8138f233-84b3-454c-a61a-4de873fa461b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1276.070489] env[61985]: DEBUG nova.compute.manager [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1276.071032] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1276.071275] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1276.071569] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1276.071647] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1276.071765] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1276.071918] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1276.072134] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1276.072301] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1276.072527] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1276.072635] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1276.072811] env[61985]: DEBUG nova.virt.hardware [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1276.073677] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff55ec9b-4384-4d31-987b-2d2f54f9e9a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.078377] env[61985]: DEBUG nova.network.neutron [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance_info_cache with network_info: [{"id": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "address": "fa:16:3e:56:1d:3d", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380110e4-5f", "ovs_interfaceid": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1276.082499] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f221a545-9058-4bac-b556-65c99732e83c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.549291] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Releasing lock "refresh_cache-d74cbed6-8f70-4e8d-b050-23e00ac334d4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1276.549618] env[61985]: DEBUG nova.compute.manager [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Instance network_info: |[{"id": "8138f233-84b3-454c-a61a-4de873fa461b", "address": "fa:16:3e:b4:0f:de", "network": {"id": "9240338e-ddc4-4cce-b5b5-bb259be32858", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-565887658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e55fe9a13574d95828bbf020d7ee62e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8138f233-84", "ovs_interfaceid": "8138f233-84b3-454c-a61a-4de873fa461b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1276.550088] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:0f:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0636c3f6-fcb7-4954-ab07-c5cd0dee37b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8138f233-84b3-454c-a61a-4de873fa461b', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1276.557444] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Creating folder: Project (2e55fe9a13574d95828bbf020d7ee62e). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1276.557720] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ad68846-c749-48f7-9efb-6a22a784fad8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.571557] env[61985]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1276.571703] env[61985]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61985) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1276.572010] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Folder already exists: Project (2e55fe9a13574d95828bbf020d7ee62e). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1276.572211] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Creating folder: Instances. Parent ref: group-v211515. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1276.572434] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca87a676-e799-402b-adb9-4e06f92a46d4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.581868] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1276.583167] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Created folder: Instances in parent group-v211515. [ 1276.583388] env[61985]: DEBUG oslo.service.loopingcall [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1276.583756] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1276.583950] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ebd88c5a-6e48-43c9-a7b0-255d6b310fd4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.603045] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1276.603045] env[61985]: value = "task-936844" [ 1276.603045] env[61985]: _type = "Task" [ 1276.603045] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.610237] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936844, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.758726] env[61985]: DEBUG nova.compute.manager [req-c07ab408-b479-4969-90d0-85ef7a493039 req-996efef7-3ae7-4dc2-a07a-974e0de8d185 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Received event network-changed-8138f233-84b3-454c-a61a-4de873fa461b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1276.758983] env[61985]: DEBUG nova.compute.manager [req-c07ab408-b479-4969-90d0-85ef7a493039 req-996efef7-3ae7-4dc2-a07a-974e0de8d185 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Refreshing instance network info cache due to event network-changed-8138f233-84b3-454c-a61a-4de873fa461b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1276.759226] env[61985]: DEBUG oslo_concurrency.lockutils [req-c07ab408-b479-4969-90d0-85ef7a493039 req-996efef7-3ae7-4dc2-a07a-974e0de8d185 service nova] Acquiring lock "refresh_cache-d74cbed6-8f70-4e8d-b050-23e00ac334d4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1276.759377] env[61985]: DEBUG oslo_concurrency.lockutils [req-c07ab408-b479-4969-90d0-85ef7a493039 req-996efef7-3ae7-4dc2-a07a-974e0de8d185 service nova] Acquired lock "refresh_cache-d74cbed6-8f70-4e8d-b050-23e00ac334d4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1276.759544] env[61985]: DEBUG nova.network.neutron [req-c07ab408-b479-4969-90d0-85ef7a493039 req-996efef7-3ae7-4dc2-a07a-974e0de8d185 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Refreshing network info cache for port 8138f233-84b3-454c-a61a-4de873fa461b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1277.085755] env[61985]: DEBUG nova.compute.manager [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61985) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1277.086058] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.086255] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.113352] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936844, 'name': CreateVM_Task, 'duration_secs': 0.453684} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.113522] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1277.114235] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211518', 'volume_id': 'a1fed62e-6782-4ff4-b5de-21f1669862fc', 'name': 'volume-a1fed62e-6782-4ff4-b5de-21f1669862fc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd74cbed6-8f70-4e8d-b050-23e00ac334d4', 'attached_at': '', 'detached_at': '', 'volume_id': 'a1fed62e-6782-4ff4-b5de-21f1669862fc', 'serial': 'a1fed62e-6782-4ff4-b5de-21f1669862fc'}, 'device_type': None, 'attachment_id': '88eab190-09d2-431d-b7e7-a9742ab07235', 'mount_device': '/dev/sda', 'disk_bus': None, 'delete_on_termination': True, 'boot_index': 0, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=61985) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1277.114455] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Root volume attach. Driver type: vmdk {{(pid=61985) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1277.115241] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401a5b5f-d7a6-41d2-a08a-bc93700c03a7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.123067] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21ad162-13b4-4dfb-b5b8-0790e6f0c410 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.129260] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2c9062-7572-4891-b5b2-78ce231ef436 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.135179] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-2a4073d2-b702-495f-8145-412c20f40831 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.142218] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for the task: (returnval){ [ 1277.142218] env[61985]: value = "task-936845" [ 1277.142218] env[61985]: _type = "Task" [ 1277.142218] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.151049] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936845, 'name': RelocateVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.447890] env[61985]: DEBUG nova.network.neutron [req-c07ab408-b479-4969-90d0-85ef7a493039 req-996efef7-3ae7-4dc2-a07a-974e0de8d185 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Updated VIF entry in instance network info cache for port 8138f233-84b3-454c-a61a-4de873fa461b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1277.448469] env[61985]: DEBUG nova.network.neutron [req-c07ab408-b479-4969-90d0-85ef7a493039 req-996efef7-3ae7-4dc2-a07a-974e0de8d185 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Updating instance_info_cache with network_info: [{"id": "8138f233-84b3-454c-a61a-4de873fa461b", "address": "fa:16:3e:b4:0f:de", "network": {"id": "9240338e-ddc4-4cce-b5b5-bb259be32858", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-565887658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e55fe9a13574d95828bbf020d7ee62e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8138f233-84", "ovs_interfaceid": "8138f233-84b3-454c-a61a-4de873fa461b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1277.589108] env[61985]: DEBUG nova.objects.instance [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'migration_context' on Instance uuid aa2eb44b-6c7c-41de-b9c2-0b01c2efe056 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1277.651441] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936845, 'name': RelocateVM_Task, 'duration_secs': 0.321739} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.651714] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Volume attach. Driver type: vmdk {{(pid=61985) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1277.651920] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211518', 'volume_id': 'a1fed62e-6782-4ff4-b5de-21f1669862fc', 'name': 'volume-a1fed62e-6782-4ff4-b5de-21f1669862fc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd74cbed6-8f70-4e8d-b050-23e00ac334d4', 'attached_at': '', 'detached_at': '', 'volume_id': 'a1fed62e-6782-4ff4-b5de-21f1669862fc', 'serial': 'a1fed62e-6782-4ff4-b5de-21f1669862fc'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1277.652680] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f656c32-7ed2-40a6-942d-9d68e2124c83 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.667847] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f70d9f-3f27-437f-8355-523ce2a7626f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.689113] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] volume-a1fed62e-6782-4ff4-b5de-21f1669862fc/volume-a1fed62e-6782-4ff4-b5de-21f1669862fc.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1277.689339] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-439c5112-8a74-4a9d-874c-7ede2ee97d55 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.708574] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for the task: (returnval){ [ 1277.708574] env[61985]: value = "task-936846" [ 1277.708574] env[61985]: _type = "Task" [ 1277.708574] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.718727] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936846, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.950994] env[61985]: DEBUG oslo_concurrency.lockutils [req-c07ab408-b479-4969-90d0-85ef7a493039 req-996efef7-3ae7-4dc2-a07a-974e0de8d185 service nova] Releasing lock "refresh_cache-d74cbed6-8f70-4e8d-b050-23e00ac334d4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1278.155670] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7528b786-e8fa-48f8-9046-c3d2b13ded2a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.163618] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45f6009-18cc-4c51-a0b6-e78c10782ff5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.194248] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58b5774-f8d0-4f6a-8b10-1490bee3174c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.201376] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd88a8d-d9bb-49cb-9b1a-0db816976078 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.214119] env[61985]: DEBUG nova.compute.provider_tree [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1278.222348] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936846, 'name': ReconfigVM_Task, 'duration_secs': 0.244036} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.223185] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Reconfigured VM instance instance-00000075 to attach disk [datastore1] volume-a1fed62e-6782-4ff4-b5de-21f1669862fc/volume-a1fed62e-6782-4ff4-b5de-21f1669862fc.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1278.228099] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10d6abdf-409c-44b5-a114-e5fe3df81049 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.244410] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for the task: (returnval){ [ 1278.244410] env[61985]: value = "task-936847" [ 1278.244410] env[61985]: _type = "Task" [ 1278.244410] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.252985] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936847, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.719160] env[61985]: DEBUG nova.scheduler.client.report [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1278.753931] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936847, 'name': ReconfigVM_Task, 'duration_secs': 0.114184} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.754256] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211518', 'volume_id': 'a1fed62e-6782-4ff4-b5de-21f1669862fc', 'name': 'volume-a1fed62e-6782-4ff4-b5de-21f1669862fc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd74cbed6-8f70-4e8d-b050-23e00ac334d4', 'attached_at': '', 'detached_at': '', 'volume_id': 'a1fed62e-6782-4ff4-b5de-21f1669862fc', 'serial': 'a1fed62e-6782-4ff4-b5de-21f1669862fc'} {{(pid=61985) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1278.754773] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a10b3bbb-8d01-4be2-95b5-d0efa17fd47f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.760177] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for the task: (returnval){ [ 1278.760177] env[61985]: value = "task-936848" [ 1278.760177] env[61985]: _type = "Task" [ 1278.760177] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.767634] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936848, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.199936] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.270233] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936848, 'name': Rename_Task, 'duration_secs': 0.116701} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.270518] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1279.270776] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e440de79-df7c-4b33-b7df-0dc21855a3c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.280217] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for the task: (returnval){ [ 1279.280217] env[61985]: value = "task-936849" [ 1279.280217] env[61985]: _type = "Task" [ 1279.280217] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.294443] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936849, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.705351] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.705554] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1279.705620] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Rebuilding the list of instances to heal {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9936}} [ 1279.729620] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.643s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.789836] env[61985]: DEBUG oslo_vmware.api [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936849, 'name': PowerOnVM_Task, 'duration_secs': 0.432879} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.790110] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1279.790322] env[61985]: INFO nova.compute.manager [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Took 3.72 seconds to spawn the instance on the hypervisor. [ 1279.790506] env[61985]: DEBUG nova.compute.manager [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1279.791279] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fe7da9-85c2-474c-b4c5-e14a78943d74 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.209748] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Skipping network cache update for instance because it is Building. {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1280.210019] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1280.210123] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquired lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.210263] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Forcefully refreshing network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1280.210410] env[61985]: DEBUG nova.objects.instance [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lazy-loading 'info_cache' on Instance uuid aa2eb44b-6c7c-41de-b9c2-0b01c2efe056 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1280.306669] env[61985]: INFO nova.compute.manager [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Took 10.80 seconds to build instance. [ 1280.808964] env[61985]: DEBUG oslo_concurrency.lockutils [None req-9673609a-4ffa-4939-a355-115ab73c78c6 tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.309s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.947246] env[61985]: DEBUG nova.compute.manager [req-9c9cb467-f24c-45bf-8dae-150d820eb359 req-5ff403a2-116e-44bd-b603-c8de67d42142 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Received event network-changed-8138f233-84b3-454c-a61a-4de873fa461b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1280.947449] env[61985]: DEBUG nova.compute.manager [req-9c9cb467-f24c-45bf-8dae-150d820eb359 req-5ff403a2-116e-44bd-b603-c8de67d42142 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Refreshing instance network info cache due to event network-changed-8138f233-84b3-454c-a61a-4de873fa461b. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1280.947672] env[61985]: DEBUG oslo_concurrency.lockutils [req-9c9cb467-f24c-45bf-8dae-150d820eb359 req-5ff403a2-116e-44bd-b603-c8de67d42142 service nova] Acquiring lock "refresh_cache-d74cbed6-8f70-4e8d-b050-23e00ac334d4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1280.947908] env[61985]: DEBUG oslo_concurrency.lockutils [req-9c9cb467-f24c-45bf-8dae-150d820eb359 req-5ff403a2-116e-44bd-b603-c8de67d42142 service nova] Acquired lock "refresh_cache-d74cbed6-8f70-4e8d-b050-23e00ac334d4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.948114] env[61985]: DEBUG nova.network.neutron [req-9c9cb467-f24c-45bf-8dae-150d820eb359 req-5ff403a2-116e-44bd-b603-c8de67d42142 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Refreshing network info cache for port 8138f233-84b3-454c-a61a-4de873fa461b {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1281.268579] env[61985]: INFO nova.compute.manager [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Swapping old allocation on dict_keys(['aed7e5b3-c662-4538-8447-c4f67b460215']) held by migration b3baf538-5899-47cc-9e45-8f0fe1e077ee for instance [ 1281.288926] env[61985]: DEBUG nova.scheduler.client.report [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Overwriting current allocation {'allocations': {'aed7e5b3-c662-4538-8447-c4f67b460215': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 168}}, 'project_id': '23a4809e353a4ecaa1bd53612fb4afc5', 'user_id': '905fa37e1f784a56b3edb9818b03d685', 'consumer_generation': 1} on consumer aa2eb44b-6c7c-41de-b9c2-0b01c2efe056 {{(pid=61985) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1281.366926] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1281.652849] env[61985]: DEBUG nova.network.neutron [req-9c9cb467-f24c-45bf-8dae-150d820eb359 req-5ff403a2-116e-44bd-b603-c8de67d42142 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Updated VIF entry in instance network info cache for port 8138f233-84b3-454c-a61a-4de873fa461b. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1281.653243] env[61985]: DEBUG nova.network.neutron [req-9c9cb467-f24c-45bf-8dae-150d820eb359 req-5ff403a2-116e-44bd-b603-c8de67d42142 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Updating instance_info_cache with network_info: [{"id": "8138f233-84b3-454c-a61a-4de873fa461b", "address": "fa:16:3e:b4:0f:de", "network": {"id": "9240338e-ddc4-4cce-b5b5-bb259be32858", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-565887658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e55fe9a13574d95828bbf020d7ee62e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8138f233-84", "ovs_interfaceid": "8138f233-84b3-454c-a61a-4de873fa461b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1281.935194] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance_info_cache with network_info: [{"id": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "address": "fa:16:3e:56:1d:3d", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380110e4-5f", "ovs_interfaceid": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1282.155519] env[61985]: DEBUG oslo_concurrency.lockutils [req-9c9cb467-f24c-45bf-8dae-150d820eb359 req-5ff403a2-116e-44bd-b603-c8de67d42142 service nova] Releasing lock "refresh_cache-d74cbed6-8f70-4e8d-b050-23e00ac334d4" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1282.438519] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Releasing lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1282.438954] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updated the network info_cache for instance {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10003}} [ 1282.439107] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.439298] env[61985]: DEBUG nova.network.neutron [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1282.440535] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1282.441217] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1282.441217] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1282.441387] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1282.944683] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.944924] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.945139] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1282.945269] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1282.946161] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3b292d-535e-4606-9825-c2b6586c37d8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.954818] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b405399b-5a70-4dda-92a1-f56d5bf17dd6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.968875] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f39e0a-d15a-47b0-8412-2aed98f78462 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.977354] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c771c3c-7f9f-459f-878c-2e5b0490e6aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.006727] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181172MB free_disk=43GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1283.006890] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1283.007079] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1283.165519] env[61985]: DEBUG nova.network.neutron [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance_info_cache with network_info: [{"id": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "address": "fa:16:3e:56:1d:3d", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap380110e4-5f", "ovs_interfaceid": "380110e4-5fc3-4dbf-b9ba-c94f348c52d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1283.668892] env[61985]: DEBUG oslo_concurrency.lockutils [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1283.669453] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1283.669688] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09cc83e6-cb11-491f-ab7e-3b1db62dc8bc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.677742] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1283.677742] env[61985]: value = "task-936850" [ 1283.677742] env[61985]: _type = "Task" [ 1283.677742] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1283.686464] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936850, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.028579] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8c95afa2-dbec-4408-a27d-595431e09748 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1284.028752] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance d74cbed6-8f70-4e8d-b050-23e00ac334d4 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1284.028892] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance aa2eb44b-6c7c-41de-b9c2-0b01c2efe056 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1284.029080] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1284.029217] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1284.071244] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4287a7e-64f0-4f95-95f8-23f76d20076f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.078655] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244eb1f1-7ac1-42d5-ae99-c0992cf379dc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.108274] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b110a507-2c5d-4509-b7e3-9a9c101e97cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.115920] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b5cf07-bcf6-4faa-92ce-14bc80036269 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.129250] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1284.186673] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936850, 'name': PowerOffVM_Task, 'duration_secs': 0.215891} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.186921] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1284.187592] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1284.187843] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1284.188021] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1284.188225] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1284.188375] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1284.188530] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1284.188752] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1284.188938] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1284.189132] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1284.189303] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1284.189480] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1284.194312] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a783b176-7362-4d98-be6b-b4efa8501405 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.209384] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1284.209384] env[61985]: value = "task-936851" [ 1284.209384] env[61985]: _type = "Task" [ 1284.209384] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.216750] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936851, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1284.632445] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1284.719493] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936851, 'name': ReconfigVM_Task, 'duration_secs': 0.144279} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1284.720258] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be4dc68-ee94-44bf-bd15-8e45ad9eb067 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.739179] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1284.739417] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1284.739578] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1284.739768] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1284.739923] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1284.740114] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1284.740339] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1284.740507] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1284.740681] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1284.740847] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1284.741041] env[61985]: DEBUG nova.virt.hardware [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1284.741790] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00a9eb19-eeaa-4db1-b2dc-28fa6ef72ffc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.746805] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1284.746805] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ae3424-6b13-bb53-8ac6-df05678f97ad" [ 1284.746805] env[61985]: _type = "Task" [ 1284.746805] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1284.754027] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ae3424-6b13-bb53-8ac6-df05678f97ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.136865] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1285.137090] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.130s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1285.257513] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ae3424-6b13-bb53-8ac6-df05678f97ad, 'name': SearchDatastore_Task, 'duration_secs': 0.00957} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.262700] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Reconfiguring VM instance instance-00000073 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1285.262972] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ce2fcc6-b0c3-4ca7-a9bb-9d04be343db5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.281144] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1285.281144] env[61985]: value = "task-936852" [ 1285.281144] env[61985]: _type = "Task" [ 1285.281144] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.288580] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936852, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.790684] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936852, 'name': ReconfigVM_Task, 'duration_secs': 0.179855} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.791180] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Reconfigured VM instance instance-00000073 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1285.791726] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce74e05-6a4b-4cf5-8c46-2d48f5e24e05 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.812946] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] aa2eb44b-6c7c-41de-b9c2-0b01c2efe056/aa2eb44b-6c7c-41de-b9c2-0b01c2efe056.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1285.813191] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-686f76ef-88a6-43ed-bf66-852029a69e33 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.830840] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1285.830840] env[61985]: value = "task-936853" [ 1285.830840] env[61985]: _type = "Task" [ 1285.830840] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.838058] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936853, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.342080] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936853, 'name': ReconfigVM_Task, 'duration_secs': 0.244197} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.342080] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Reconfigured VM instance instance-00000073 to attach disk [datastore1] aa2eb44b-6c7c-41de-b9c2-0b01c2efe056/aa2eb44b-6c7c-41de-b9c2-0b01c2efe056.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1286.342446] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df61cf5b-3b80-4355-87bc-aadae177cdd0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.360248] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92647d09-2421-4a57-86fc-565a4e33583d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.378325] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e41c3cf-d237-48b2-b0d3-52a7e8cb6e4c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.395145] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2313c0c6-1e3f-42ac-a510-94e3d2ef5a3a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.401285] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1286.401495] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7fa3e25-a32b-45ce-863f-7b0177a21d96 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.407132] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1286.407132] env[61985]: value = "task-936854" [ 1286.407132] env[61985]: _type = "Task" [ 1286.407132] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.413985] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936854, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.918291] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936854, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.417889] env[61985]: DEBUG oslo_vmware.api [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936854, 'name': PowerOnVM_Task, 'duration_secs': 0.538956} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.418177] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1288.429607] env[61985]: INFO nova.compute.manager [None req-dcf8a713-d97b-4dd9-b62e-02040d78f885 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance to original state: 'active' [ 1290.223082] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "8c95afa2-dbec-4408-a27d-595431e09748" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1290.223475] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "8c95afa2-dbec-4408-a27d-595431e09748" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1290.223608] env[61985]: INFO nova.compute.manager [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Shelving [ 1290.357696] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1290.358040] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1290.358262] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1290.358451] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1290.358629] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1290.360763] env[61985]: INFO nova.compute.manager [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Terminating instance [ 1290.362716] env[61985]: DEBUG nova.compute.manager [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1290.362958] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1290.363950] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bf4b54-bf1f-4f64-8987-c8c21f22a94a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.372256] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1290.372482] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49776668-9f62-4d23-9e4a-8786a21a86e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.379765] env[61985]: DEBUG oslo_vmware.api [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1290.379765] env[61985]: value = "task-936855" [ 1290.379765] env[61985]: _type = "Task" [ 1290.379765] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.388202] env[61985]: DEBUG oslo_vmware.api [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936855, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.731524] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1290.731784] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55ff0866-f7a1-405b-b51d-df3aa7a4ac17 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.739381] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1290.739381] env[61985]: value = "task-936856" [ 1290.739381] env[61985]: _type = "Task" [ 1290.739381] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.747360] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936856, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.890204] env[61985]: DEBUG oslo_vmware.api [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936855, 'name': PowerOffVM_Task, 'duration_secs': 0.275251} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.890498] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1290.890673] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1290.890960] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79f43bd8-1a24-45a0-a07c-4b0083045c61 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.955424] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1290.955636] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1290.955873] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleting the datastore file [datastore1] aa2eb44b-6c7c-41de-b9c2-0b01c2efe056 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1290.956167] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ba9f9b1-e346-44a9-a141-5ef095ac51ab {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.962047] env[61985]: DEBUG oslo_vmware.api [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1290.962047] env[61985]: value = "task-936858" [ 1290.962047] env[61985]: _type = "Task" [ 1290.962047] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.969354] env[61985]: DEBUG oslo_vmware.api [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936858, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.249876] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936856, 'name': PowerOffVM_Task, 'duration_secs': 0.148869} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.250172] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1291.250910] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a31daea-b2dd-4528-b8f7-330ae69b6f96 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.268282] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcfeab39-6254-4aa9-af8d-59d227d9ad4d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.471161] env[61985]: DEBUG oslo_vmware.api [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936858, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128799} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.471394] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1291.471564] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1291.471746] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1291.471929] env[61985]: INFO nova.compute.manager [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1291.472191] env[61985]: DEBUG oslo.service.loopingcall [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1291.472389] env[61985]: DEBUG nova.compute.manager [-] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1291.472482] env[61985]: DEBUG nova.network.neutron [-] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1291.777757] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Creating Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1291.778105] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e74b87b0-264d-474b-8bf3-c8b8e064254e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.785813] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1291.785813] env[61985]: value = "task-936859" [ 1291.785813] env[61985]: _type = "Task" [ 1291.785813] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.794176] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936859, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.896399] env[61985]: DEBUG nova.compute.manager [req-09892a88-2abe-4686-a9ce-613c13d7d916 req-3aaa0a6d-1a6f-4f36-a1c9-cf3df54e16c0 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Received event network-vif-deleted-380110e4-5fc3-4dbf-b9ba-c94f348c52d2 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1291.896480] env[61985]: INFO nova.compute.manager [req-09892a88-2abe-4686-a9ce-613c13d7d916 req-3aaa0a6d-1a6f-4f36-a1c9-cf3df54e16c0 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Neutron deleted interface 380110e4-5fc3-4dbf-b9ba-c94f348c52d2; detaching it from the instance and deleting it from the info cache [ 1291.896677] env[61985]: DEBUG nova.network.neutron [req-09892a88-2abe-4686-a9ce-613c13d7d916 req-3aaa0a6d-1a6f-4f36-a1c9-cf3df54e16c0 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1292.295746] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936859, 'name': CreateSnapshot_Task, 'duration_secs': 0.429423} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.296069] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Created Snapshot of the VM instance {{(pid=61985) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1292.296763] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13726e79-c600-4f32-b78b-ab6d3486dad9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.368705] env[61985]: DEBUG nova.network.neutron [-] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1292.400246] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe73f8be-ea95-4b4e-a7b9-e5091b5c344e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.409562] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b80d2c-71e2-4115-8e86-bb9a31259e6b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.433182] env[61985]: DEBUG nova.compute.manager [req-09892a88-2abe-4686-a9ce-613c13d7d916 req-3aaa0a6d-1a6f-4f36-a1c9-cf3df54e16c0 service nova] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Detach interface failed, port_id=380110e4-5fc3-4dbf-b9ba-c94f348c52d2, reason: Instance aa2eb44b-6c7c-41de-b9c2-0b01c2efe056 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1292.813377] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Creating linked-clone VM from snapshot {{(pid=61985) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1292.813714] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-99c36953-4e2e-4195-9f80-729c262e5718 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.822160] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1292.822160] env[61985]: value = "task-936860" [ 1292.822160] env[61985]: _type = "Task" [ 1292.822160] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.829725] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936860, 'name': CloneVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.871414] env[61985]: INFO nova.compute.manager [-] [instance: aa2eb44b-6c7c-41de-b9c2-0b01c2efe056] Took 1.40 seconds to deallocate network for instance. [ 1293.332861] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936860, 'name': CloneVM_Task} progress is 94%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.378195] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1293.378486] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1293.378765] env[61985]: DEBUG nova.objects.instance [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'resources' on Instance uuid aa2eb44b-6c7c-41de-b9c2-0b01c2efe056 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1293.832961] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936860, 'name': CloneVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.932431] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0003868b-3290-494a-9b53-4741bc8bb21e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.940589] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea68b130-c9a1-4b2b-a617-c8fb2dcf2520 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.971594] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9e11e9-e48d-40f6-b32d-468975ed3672 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.978658] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a979f0b-46e6-4ea2-b32a-5116b6bd5731 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.991638] env[61985]: DEBUG nova.compute.provider_tree [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1294.333988] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936860, 'name': CloneVM_Task, 'duration_secs': 1.022002} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.334306] env[61985]: INFO nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Created linked-clone VM from snapshot [ 1294.335091] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e29c3f-e97d-4f2c-9fdc-585162af6f71 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.342285] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Uploading image c4228803-7851-47f8-b83e-fc58a07898d8 {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1294.366572] env[61985]: DEBUG oslo_vmware.rw_handles [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1294.366572] env[61985]: value = "vm-211522" [ 1294.366572] env[61985]: _type = "VirtualMachine" [ 1294.366572] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1294.366822] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f280abec-4bc8-4d31-8995-038a66ca61dd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.373759] env[61985]: DEBUG oslo_vmware.rw_handles [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lease: (returnval){ [ 1294.373759] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c0c754-3d05-8bfe-6454-68d0fc1f9c29" [ 1294.373759] env[61985]: _type = "HttpNfcLease" [ 1294.373759] env[61985]: } obtained for exporting VM: (result){ [ 1294.373759] env[61985]: value = "vm-211522" [ 1294.373759] env[61985]: _type = "VirtualMachine" [ 1294.373759] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1294.374194] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the lease: (returnval){ [ 1294.374194] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c0c754-3d05-8bfe-6454-68d0fc1f9c29" [ 1294.374194] env[61985]: _type = "HttpNfcLease" [ 1294.374194] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1294.380043] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1294.380043] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c0c754-3d05-8bfe-6454-68d0fc1f9c29" [ 1294.380043] env[61985]: _type = "HttpNfcLease" [ 1294.380043] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1294.494433] env[61985]: DEBUG nova.scheduler.client.report [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1294.882582] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1294.882582] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c0c754-3d05-8bfe-6454-68d0fc1f9c29" [ 1294.882582] env[61985]: _type = "HttpNfcLease" [ 1294.882582] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1294.882941] env[61985]: DEBUG oslo_vmware.rw_handles [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1294.882941] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52c0c754-3d05-8bfe-6454-68d0fc1f9c29" [ 1294.882941] env[61985]: _type = "HttpNfcLease" [ 1294.882941] env[61985]: }. {{(pid=61985) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1294.883525] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ba1def-700a-40da-bd61-658d18b4bda3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.890146] env[61985]: DEBUG oslo_vmware.rw_handles [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da1cca-66bb-d857-4d58-4a81bcbd8820/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1294.890322] env[61985]: DEBUG oslo_vmware.rw_handles [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da1cca-66bb-d857-4d58-4a81bcbd8820/disk-0.vmdk for reading. {{(pid=61985) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1294.974678] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-12f7c76c-3c2d-46de-8c40-81c038420f18 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.002256] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.623s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.018974] env[61985]: INFO nova.scheduler.client.report [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleted allocations for instance aa2eb44b-6c7c-41de-b9c2-0b01c2efe056 [ 1295.526792] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4bfb017b-43fc-40a2-b355-f43cfdb7822f tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "aa2eb44b-6c7c-41de-b9c2-0b01c2efe056" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.169s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1296.398107] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "4908dca0-b19f-4fff-b560-075ad46020d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1296.399352] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "4908dca0-b19f-4fff-b560-075ad46020d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1296.900873] env[61985]: DEBUG nova.compute.manager [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1297.422824] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1297.423181] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1297.424714] env[61985]: INFO nova.compute.claims [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1298.485913] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81076ece-5912-4693-9622-785e3f77aa63 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.494305] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc22906-49e8-4a8b-a04d-f572e45d11b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.524473] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59aaa2a-c8bc-4c94-9029-c9dffc8cc293 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.532185] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12ce8bf-44da-4b94-acd2-b273e60bce7e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.546977] env[61985]: DEBUG nova.compute.provider_tree [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1299.050202] env[61985]: DEBUG nova.scheduler.client.report [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1299.555149] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.132s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1299.555761] env[61985]: DEBUG nova.compute.manager [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1300.061906] env[61985]: DEBUG nova.compute.utils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1300.063895] env[61985]: DEBUG nova.compute.manager [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1300.064099] env[61985]: DEBUG nova.network.neutron [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1300.112437] env[61985]: DEBUG nova.policy [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '905fa37e1f784a56b3edb9818b03d685', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23a4809e353a4ecaa1bd53612fb4afc5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1300.403509] env[61985]: DEBUG nova.network.neutron [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Successfully created port: 1311d04d-d434-4c2a-ac57-ce6c00626037 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1300.567941] env[61985]: DEBUG nova.compute.manager [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1301.580606] env[61985]: DEBUG nova.compute.manager [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1301.605377] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1301.605640] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1301.605835] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1301.606046] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1301.606210] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1301.606365] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1301.606581] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1301.606749] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1301.606928] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1301.607115] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1301.607301] env[61985]: DEBUG nova.virt.hardware [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1301.608225] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52eeceed-ce43-4f86-9996-3d55be67bedc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.616309] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab0f701-9227-4fff-9769-07ca654e959c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.789714] env[61985]: DEBUG nova.compute.manager [req-fbb76766-5766-4f0b-b34f-30c2ce7016e7 req-1d99311b-5116-4bab-a567-50717ee1fdfa service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Received event network-vif-plugged-1311d04d-d434-4c2a-ac57-ce6c00626037 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1301.789977] env[61985]: DEBUG oslo_concurrency.lockutils [req-fbb76766-5766-4f0b-b34f-30c2ce7016e7 req-1d99311b-5116-4bab-a567-50717ee1fdfa service nova] Acquiring lock "4908dca0-b19f-4fff-b560-075ad46020d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1301.790320] env[61985]: DEBUG oslo_concurrency.lockutils [req-fbb76766-5766-4f0b-b34f-30c2ce7016e7 req-1d99311b-5116-4bab-a567-50717ee1fdfa service nova] Lock "4908dca0-b19f-4fff-b560-075ad46020d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1301.790417] env[61985]: DEBUG oslo_concurrency.lockutils [req-fbb76766-5766-4f0b-b34f-30c2ce7016e7 req-1d99311b-5116-4bab-a567-50717ee1fdfa service nova] Lock "4908dca0-b19f-4fff-b560-075ad46020d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1301.790599] env[61985]: DEBUG nova.compute.manager [req-fbb76766-5766-4f0b-b34f-30c2ce7016e7 req-1d99311b-5116-4bab-a567-50717ee1fdfa service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] No waiting events found dispatching network-vif-plugged-1311d04d-d434-4c2a-ac57-ce6c00626037 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1301.790759] env[61985]: WARNING nova.compute.manager [req-fbb76766-5766-4f0b-b34f-30c2ce7016e7 req-1d99311b-5116-4bab-a567-50717ee1fdfa service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Received unexpected event network-vif-plugged-1311d04d-d434-4c2a-ac57-ce6c00626037 for instance with vm_state building and task_state spawning. [ 1301.882602] env[61985]: DEBUG nova.network.neutron [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Successfully updated port: 1311d04d-d434-4c2a-ac57-ce6c00626037 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1302.384918] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1302.385153] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1302.385293] env[61985]: DEBUG nova.network.neutron [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1302.697652] env[61985]: DEBUG oslo_vmware.rw_handles [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da1cca-66bb-d857-4d58-4a81bcbd8820/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1302.698668] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d483813-fe6f-4037-b8b1-e0886e65ba28 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.706676] env[61985]: DEBUG oslo_vmware.rw_handles [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da1cca-66bb-d857-4d58-4a81bcbd8820/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1302.706852] env[61985]: ERROR oslo_vmware.rw_handles [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da1cca-66bb-d857-4d58-4a81bcbd8820/disk-0.vmdk due to incomplete transfer. [ 1302.707088] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d3ae6e8d-8904-479b-9b6c-bda00b503ef8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.714048] env[61985]: DEBUG oslo_vmware.rw_handles [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da1cca-66bb-d857-4d58-4a81bcbd8820/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1302.714255] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Uploaded image c4228803-7851-47f8-b83e-fc58a07898d8 to the Glance image server {{(pid=61985) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1302.716462] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Destroying the VM {{(pid=61985) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1302.716691] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a17a7084-1511-40f6-9f59-0de2ce9dc6aa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.721778] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1302.721778] env[61985]: value = "task-936862" [ 1302.721778] env[61985]: _type = "Task" [ 1302.721778] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.728825] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936862, 'name': Destroy_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.917941] env[61985]: DEBUG nova.network.neutron [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1303.039506] env[61985]: DEBUG nova.network.neutron [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Updating instance_info_cache with network_info: [{"id": "1311d04d-d434-4c2a-ac57-ce6c00626037", "address": "fa:16:3e:db:23:37", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1311d04d-d4", "ovs_interfaceid": "1311d04d-d434-4c2a-ac57-ce6c00626037", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1303.238864] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936862, 'name': Destroy_Task, 'duration_secs': 0.36877} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.239157] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Destroyed the VM [ 1303.239403] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Deleting Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1303.239702] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-97a15544-0e61-425a-9902-10b55789571d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.245452] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1303.245452] env[61985]: value = "task-936863" [ 1303.245452] env[61985]: _type = "Task" [ 1303.245452] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.253084] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936863, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.542593] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1303.542957] env[61985]: DEBUG nova.compute.manager [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Instance network_info: |[{"id": "1311d04d-d434-4c2a-ac57-ce6c00626037", "address": "fa:16:3e:db:23:37", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1311d04d-d4", "ovs_interfaceid": "1311d04d-d434-4c2a-ac57-ce6c00626037", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1303.543450] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:23:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1311d04d-d434-4c2a-ac57-ce6c00626037', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1303.550968] env[61985]: DEBUG oslo.service.loopingcall [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1303.551239] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1303.551467] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c402933f-4a0e-4253-837d-cb52855e9bfc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.571353] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1303.571353] env[61985]: value = "task-936864" [ 1303.571353] env[61985]: _type = "Task" [ 1303.571353] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.580411] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936864, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.755951] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936863, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.816246] env[61985]: DEBUG nova.compute.manager [req-2ac3c0dd-4769-4ccf-b394-e4fb2548bccc req-0b2e3ce8-1ae9-4257-8f47-88fb57552069 service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Received event network-changed-1311d04d-d434-4c2a-ac57-ce6c00626037 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1303.816689] env[61985]: DEBUG nova.compute.manager [req-2ac3c0dd-4769-4ccf-b394-e4fb2548bccc req-0b2e3ce8-1ae9-4257-8f47-88fb57552069 service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Refreshing instance network info cache due to event network-changed-1311d04d-d434-4c2a-ac57-ce6c00626037. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1303.816996] env[61985]: DEBUG oslo_concurrency.lockutils [req-2ac3c0dd-4769-4ccf-b394-e4fb2548bccc req-0b2e3ce8-1ae9-4257-8f47-88fb57552069 service nova] Acquiring lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1303.817201] env[61985]: DEBUG oslo_concurrency.lockutils [req-2ac3c0dd-4769-4ccf-b394-e4fb2548bccc req-0b2e3ce8-1ae9-4257-8f47-88fb57552069 service nova] Acquired lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1303.817383] env[61985]: DEBUG nova.network.neutron [req-2ac3c0dd-4769-4ccf-b394-e4fb2548bccc req-0b2e3ce8-1ae9-4257-8f47-88fb57552069 service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Refreshing network info cache for port 1311d04d-d434-4c2a-ac57-ce6c00626037 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1304.082602] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936864, 'name': CreateVM_Task, 'duration_secs': 0.326345} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.082803] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1304.083418] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1304.083591] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1304.083926] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1304.084232] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4010390a-9daa-4b2f-a829-3d04467caebc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.089122] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1304.089122] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e6a82b-6f7e-79cb-b5be-cf2ba74e85c7" [ 1304.089122] env[61985]: _type = "Task" [ 1304.089122] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.098237] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e6a82b-6f7e-79cb-b5be-cf2ba74e85c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.256019] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936863, 'name': RemoveSnapshot_Task, 'duration_secs': 0.665251} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.256275] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Deleted Snapshot of the VM instance {{(pid=61985) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1304.256543] env[61985]: DEBUG nova.compute.manager [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1304.257290] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31438461-1af9-477b-a494-2b979f3dc55b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.503876] env[61985]: DEBUG nova.network.neutron [req-2ac3c0dd-4769-4ccf-b394-e4fb2548bccc req-0b2e3ce8-1ae9-4257-8f47-88fb57552069 service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Updated VIF entry in instance network info cache for port 1311d04d-d434-4c2a-ac57-ce6c00626037. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1304.504408] env[61985]: DEBUG nova.network.neutron [req-2ac3c0dd-4769-4ccf-b394-e4fb2548bccc req-0b2e3ce8-1ae9-4257-8f47-88fb57552069 service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Updating instance_info_cache with network_info: [{"id": "1311d04d-d434-4c2a-ac57-ce6c00626037", "address": "fa:16:3e:db:23:37", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1311d04d-d4", "ovs_interfaceid": "1311d04d-d434-4c2a-ac57-ce6c00626037", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1304.599272] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52e6a82b-6f7e-79cb-b5be-cf2ba74e85c7, 'name': SearchDatastore_Task, 'duration_secs': 0.010614} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.599516] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1304.599744] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1304.599981] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1304.600152] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1304.600368] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1304.600637] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b10d916b-f7b0-4d2f-a295-da85d94b0878 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.608279] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1304.608512] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1304.609147] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0005dee3-15cc-415b-aff8-f7e8a61abd79 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.615209] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1304.615209] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ec085a-b7d2-aa8c-f7a6-99de05bb2d61" [ 1304.615209] env[61985]: _type = "Task" [ 1304.615209] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.622307] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ec085a-b7d2-aa8c-f7a6-99de05bb2d61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.769479] env[61985]: INFO nova.compute.manager [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Shelve offloading [ 1304.771153] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1304.771641] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5acda394-88b3-4f5a-8d8c-15dc43603454 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.779020] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1304.779020] env[61985]: value = "task-936865" [ 1304.779020] env[61985]: _type = "Task" [ 1304.779020] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.786047] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.007768] env[61985]: DEBUG oslo_concurrency.lockutils [req-2ac3c0dd-4769-4ccf-b394-e4fb2548bccc req-0b2e3ce8-1ae9-4257-8f47-88fb57552069 service nova] Releasing lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1305.125605] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52ec085a-b7d2-aa8c-f7a6-99de05bb2d61, 'name': SearchDatastore_Task, 'duration_secs': 0.007967} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.126380] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-471b1eb9-9164-45ec-b459-262647e56cc7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.131233] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1305.131233] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528dbdc8-e944-2ade-28aa-9fde69e88cf2" [ 1305.131233] env[61985]: _type = "Task" [ 1305.131233] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.138456] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528dbdc8-e944-2ade-28aa-9fde69e88cf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.288852] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1305.289210] env[61985]: DEBUG nova.compute.manager [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1305.289945] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b841db-75d7-4336-9acc-61c977f18bf3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.295989] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1305.296175] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1305.296349] env[61985]: DEBUG nova.network.neutron [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1305.641663] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]528dbdc8-e944-2ade-28aa-9fde69e88cf2, 'name': SearchDatastore_Task, 'duration_secs': 0.008884} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.641940] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1305.642228] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 4908dca0-b19f-4fff-b560-075ad46020d3/4908dca0-b19f-4fff-b560-075ad46020d3.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1305.642477] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ecd1697-58ac-4ed8-bed8-003a343513a3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.649199] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1305.649199] env[61985]: value = "task-936866" [ 1305.649199] env[61985]: _type = "Task" [ 1305.649199] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.655932] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936866, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.049915] env[61985]: DEBUG nova.network.neutron [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating instance_info_cache with network_info: [{"id": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "address": "fa:16:3e:e1:8c:42", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71d390e-e1", "ovs_interfaceid": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1306.159700] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936866, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439169} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.159956] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore1] 4908dca0-b19f-4fff-b560-075ad46020d3/4908dca0-b19f-4fff-b560-075ad46020d3.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1306.160232] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1306.160481] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3595bc2d-6228-4725-9ae9-94ee5e4176e0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.167223] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1306.167223] env[61985]: value = "task-936867" [ 1306.167223] env[61985]: _type = "Task" [ 1306.167223] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.174386] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936867, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.553509] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1306.677457] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936867, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06226} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.677747] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1306.678571] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19214f77-ee46-4e4c-99eb-b9a2a4938b21 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.699966] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Reconfiguring VM instance instance-00000076 to attach disk [datastore1] 4908dca0-b19f-4fff-b560-075ad46020d3/4908dca0-b19f-4fff-b560-075ad46020d3.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1306.700406] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da269b42-03b5-4110-af1a-856dc701ba32 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.721560] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1306.721560] env[61985]: value = "task-936868" [ 1306.721560] env[61985]: _type = "Task" [ 1306.721560] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.729597] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936868, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.803086] env[61985]: DEBUG nova.compute.manager [req-5df147ff-63d1-4d13-97cd-abb03f6da5bb req-7a279ae3-2a8a-41ea-b457-566864e32793 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received event network-vif-unplugged-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1306.803355] env[61985]: DEBUG oslo_concurrency.lockutils [req-5df147ff-63d1-4d13-97cd-abb03f6da5bb req-7a279ae3-2a8a-41ea-b457-566864e32793 service nova] Acquiring lock "8c95afa2-dbec-4408-a27d-595431e09748-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1306.803621] env[61985]: DEBUG oslo_concurrency.lockutils [req-5df147ff-63d1-4d13-97cd-abb03f6da5bb req-7a279ae3-2a8a-41ea-b457-566864e32793 service nova] Lock "8c95afa2-dbec-4408-a27d-595431e09748-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1306.803820] env[61985]: DEBUG oslo_concurrency.lockutils [req-5df147ff-63d1-4d13-97cd-abb03f6da5bb req-7a279ae3-2a8a-41ea-b457-566864e32793 service nova] Lock "8c95afa2-dbec-4408-a27d-595431e09748-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1306.804055] env[61985]: DEBUG nova.compute.manager [req-5df147ff-63d1-4d13-97cd-abb03f6da5bb req-7a279ae3-2a8a-41ea-b457-566864e32793 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] No waiting events found dispatching network-vif-unplugged-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1306.804295] env[61985]: WARNING nova.compute.manager [req-5df147ff-63d1-4d13-97cd-abb03f6da5bb req-7a279ae3-2a8a-41ea-b457-566864e32793 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received unexpected event network-vif-unplugged-a71d390e-e1df-4d20-a37a-e5a2331f9b7f for instance with vm_state shelved and task_state shelving_offloading. [ 1306.936515] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1306.937491] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41708fd5-4887-4409-a701-36ae9fcd67c2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.945253] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1306.945506] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-793b8df0-4417-4af6-ae72-354ef93b5aee {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.011950] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1307.012253] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1307.012471] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleting the datastore file [datastore2] 8c95afa2-dbec-4408-a27d-595431e09748 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1307.012901] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-071ff2a6-c7c7-4405-add9-2092917dcae7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.021806] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1307.021806] env[61985]: value = "task-936870" [ 1307.021806] env[61985]: _type = "Task" [ 1307.021806] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.031559] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936870, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.231870] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936868, 'name': ReconfigVM_Task, 'duration_secs': 0.385046} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.232274] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Reconfigured VM instance instance-00000076 to attach disk [datastore1] 4908dca0-b19f-4fff-b560-075ad46020d3/4908dca0-b19f-4fff-b560-075ad46020d3.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1307.232797] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a4860d8-0552-4d32-ac34-de247de7b8ef {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.238705] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1307.238705] env[61985]: value = "task-936871" [ 1307.238705] env[61985]: _type = "Task" [ 1307.238705] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.245933] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936871, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.531655] env[61985]: DEBUG oslo_vmware.api [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936870, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164459} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.531975] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1307.532233] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1307.532467] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1307.553413] env[61985]: INFO nova.scheduler.client.report [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleted allocations for instance 8c95afa2-dbec-4408-a27d-595431e09748 [ 1307.748441] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936871, 'name': Rename_Task, 'duration_secs': 0.13851} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.748755] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1307.748985] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5d28582-2ea5-4a28-976a-8c277443e808 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.755647] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1307.755647] env[61985]: value = "task-936872" [ 1307.755647] env[61985]: _type = "Task" [ 1307.755647] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.769062] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936872, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.057692] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1308.058086] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1308.058349] env[61985]: DEBUG nova.objects.instance [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'resources' on Instance uuid 8c95afa2-dbec-4408-a27d-595431e09748 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1308.265498] env[61985]: DEBUG oslo_vmware.api [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936872, 'name': PowerOnVM_Task, 'duration_secs': 0.47085} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.265781] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1308.265974] env[61985]: INFO nova.compute.manager [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Took 6.69 seconds to spawn the instance on the hypervisor. [ 1308.266184] env[61985]: DEBUG nova.compute.manager [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1308.267050] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb46a2e6-f10d-46a1-b84f-88b345c2d68c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.561599] env[61985]: DEBUG nova.objects.instance [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'numa_topology' on Instance uuid 8c95afa2-dbec-4408-a27d-595431e09748 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1308.786844] env[61985]: INFO nova.compute.manager [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Took 11.38 seconds to build instance. [ 1308.830247] env[61985]: DEBUG nova.compute.manager [req-b19d3319-2822-4723-978a-bfae2309d723 req-2b738877-798a-460d-93f8-a6bee68e67b9 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received event network-changed-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1308.830452] env[61985]: DEBUG nova.compute.manager [req-b19d3319-2822-4723-978a-bfae2309d723 req-2b738877-798a-460d-93f8-a6bee68e67b9 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Refreshing instance network info cache due to event network-changed-a71d390e-e1df-4d20-a37a-e5a2331f9b7f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1308.830673] env[61985]: DEBUG oslo_concurrency.lockutils [req-b19d3319-2822-4723-978a-bfae2309d723 req-2b738877-798a-460d-93f8-a6bee68e67b9 service nova] Acquiring lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1308.830824] env[61985]: DEBUG oslo_concurrency.lockutils [req-b19d3319-2822-4723-978a-bfae2309d723 req-2b738877-798a-460d-93f8-a6bee68e67b9 service nova] Acquired lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1308.830990] env[61985]: DEBUG nova.network.neutron [req-b19d3319-2822-4723-978a-bfae2309d723 req-2b738877-798a-460d-93f8-a6bee68e67b9 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Refreshing network info cache for port a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1309.064380] env[61985]: DEBUG nova.objects.base [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Object Instance<8c95afa2-dbec-4408-a27d-595431e09748> lazy-loaded attributes: resources,numa_topology {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1309.112065] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27853c67-924d-4fa8-9c10-cd0544aca16b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.120419] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a426b86-a4bf-4c80-84ec-be9a95b199c0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.151822] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac30597-5fb8-46c9-897b-ca9ddcde57fa {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.159425] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81ee665-245b-4d41-99bf-eb88a9e629b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.173162] env[61985]: DEBUG nova.compute.provider_tree [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1309.288476] env[61985]: DEBUG oslo_concurrency.lockutils [None req-a2f19901-295b-4d52-9d23-964dcbb0b8a8 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "4908dca0-b19f-4fff-b560-075ad46020d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.890s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1309.676715] env[61985]: DEBUG nova.scheduler.client.report [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1309.750033] env[61985]: DEBUG nova.network.neutron [req-b19d3319-2822-4723-978a-bfae2309d723 req-2b738877-798a-460d-93f8-a6bee68e67b9 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updated VIF entry in instance network info cache for port a71d390e-e1df-4d20-a37a-e5a2331f9b7f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1309.750486] env[61985]: DEBUG nova.network.neutron [req-b19d3319-2822-4723-978a-bfae2309d723 req-2b738877-798a-460d-93f8-a6bee68e67b9 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating instance_info_cache with network_info: [{"id": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "address": "fa:16:3e:e1:8c:42", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapa71d390e-e1", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1310.161346] env[61985]: DEBUG nova.compute.manager [req-2a59ae4b-dd75-41bd-bec2-79c229f93711 req-ac285237-472e-4e5b-9530-d3548e94bea6 service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Received event network-changed-1311d04d-d434-4c2a-ac57-ce6c00626037 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1310.161558] env[61985]: DEBUG nova.compute.manager [req-2a59ae4b-dd75-41bd-bec2-79c229f93711 req-ac285237-472e-4e5b-9530-d3548e94bea6 service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Refreshing instance network info cache due to event network-changed-1311d04d-d434-4c2a-ac57-ce6c00626037. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1310.161775] env[61985]: DEBUG oslo_concurrency.lockutils [req-2a59ae4b-dd75-41bd-bec2-79c229f93711 req-ac285237-472e-4e5b-9530-d3548e94bea6 service nova] Acquiring lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1310.161924] env[61985]: DEBUG oslo_concurrency.lockutils [req-2a59ae4b-dd75-41bd-bec2-79c229f93711 req-ac285237-472e-4e5b-9530-d3548e94bea6 service nova] Acquired lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1310.162103] env[61985]: DEBUG nova.network.neutron [req-2a59ae4b-dd75-41bd-bec2-79c229f93711 req-ac285237-472e-4e5b-9530-d3548e94bea6 service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Refreshing network info cache for port 1311d04d-d434-4c2a-ac57-ce6c00626037 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1310.184642] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.127s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1310.257171] env[61985]: DEBUG oslo_concurrency.lockutils [req-b19d3319-2822-4723-978a-bfae2309d723 req-2b738877-798a-460d-93f8-a6bee68e67b9 service nova] Releasing lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1310.257901] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "8c95afa2-dbec-4408-a27d-595431e09748" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1310.693897] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5fd10520-120b-4b2c-8e71-f9692758dc8f tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "8c95afa2-dbec-4408-a27d-595431e09748" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.470s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1310.694938] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "8c95afa2-dbec-4408-a27d-595431e09748" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.437s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1310.695075] env[61985]: INFO nova.compute.manager [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Unshelving [ 1310.897398] env[61985]: DEBUG nova.network.neutron [req-2a59ae4b-dd75-41bd-bec2-79c229f93711 req-ac285237-472e-4e5b-9530-d3548e94bea6 service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Updated VIF entry in instance network info cache for port 1311d04d-d434-4c2a-ac57-ce6c00626037. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1310.897853] env[61985]: DEBUG nova.network.neutron [req-2a59ae4b-dd75-41bd-bec2-79c229f93711 req-ac285237-472e-4e5b-9530-d3548e94bea6 service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Updating instance_info_cache with network_info: [{"id": "1311d04d-d434-4c2a-ac57-ce6c00626037", "address": "fa:16:3e:db:23:37", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1311d04d-d4", "ovs_interfaceid": "1311d04d-d434-4c2a-ac57-ce6c00626037", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1311.400371] env[61985]: DEBUG oslo_concurrency.lockutils [req-2a59ae4b-dd75-41bd-bec2-79c229f93711 req-ac285237-472e-4e5b-9530-d3548e94bea6 service nova] Releasing lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1311.721384] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1311.721646] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1311.721859] env[61985]: DEBUG nova.objects.instance [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'pci_requests' on Instance uuid 8c95afa2-dbec-4408-a27d-595431e09748 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1312.225525] env[61985]: DEBUG nova.objects.instance [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'numa_topology' on Instance uuid 8c95afa2-dbec-4408-a27d-595431e09748 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1312.728484] env[61985]: INFO nova.compute.claims [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1313.789469] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0438a1ca-fe04-4a7e-8c16-e2942236c3f9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.797216] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa587e3-216a-4564-bde5-d29ff0bc20d7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.827700] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b0a5d8-eb30-4626-be85-44cdfc989012 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.834779] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8214d43-4947-433c-961d-0103069b6f8a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.847617] env[61985]: DEBUG nova.compute.provider_tree [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1314.351316] env[61985]: DEBUG nova.scheduler.client.report [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1314.856822] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.135s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1314.884109] env[61985]: INFO nova.network.neutron [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating port a71d390e-e1df-4d20-a37a-e5a2331f9b7f with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1316.776653] env[61985]: DEBUG nova.compute.manager [req-30007c38-f2d9-479c-aaa0-aad97284ab7b req-9a1f205c-6f2b-43fc-88b4-8ffd1459199f service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received event network-vif-plugged-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1316.776895] env[61985]: DEBUG oslo_concurrency.lockutils [req-30007c38-f2d9-479c-aaa0-aad97284ab7b req-9a1f205c-6f2b-43fc-88b4-8ffd1459199f service nova] Acquiring lock "8c95afa2-dbec-4408-a27d-595431e09748-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1316.777120] env[61985]: DEBUG oslo_concurrency.lockutils [req-30007c38-f2d9-479c-aaa0-aad97284ab7b req-9a1f205c-6f2b-43fc-88b4-8ffd1459199f service nova] Lock "8c95afa2-dbec-4408-a27d-595431e09748-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1316.777304] env[61985]: DEBUG oslo_concurrency.lockutils [req-30007c38-f2d9-479c-aaa0-aad97284ab7b req-9a1f205c-6f2b-43fc-88b4-8ffd1459199f service nova] Lock "8c95afa2-dbec-4408-a27d-595431e09748-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1316.777479] env[61985]: DEBUG nova.compute.manager [req-30007c38-f2d9-479c-aaa0-aad97284ab7b req-9a1f205c-6f2b-43fc-88b4-8ffd1459199f service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] No waiting events found dispatching network-vif-plugged-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1316.777651] env[61985]: WARNING nova.compute.manager [req-30007c38-f2d9-479c-aaa0-aad97284ab7b req-9a1f205c-6f2b-43fc-88b4-8ffd1459199f service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received unexpected event network-vif-plugged-a71d390e-e1df-4d20-a37a-e5a2331f9b7f for instance with vm_state shelved_offloaded and task_state spawning. [ 1316.862660] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1316.862940] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1316.863070] env[61985]: DEBUG nova.network.neutron [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1317.567206] env[61985]: DEBUG nova.network.neutron [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating instance_info_cache with network_info: [{"id": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "address": "fa:16:3e:e1:8c:42", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71d390e-e1", "ovs_interfaceid": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1318.070596] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1318.096423] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='59ffc11f8c911aee5d91c4e50be9944c',container_format='bare',created_at=2024-09-18T01:36:21Z,direct_url=,disk_format='vmdk',id=c4228803-7851-47f8-b83e-fc58a07898d8,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1472571627-shelved',owner='e7e22f7d7b7b407ebb2d00e174d8a42c',properties=ImageMetaProps,protected=,size=31666688,status='active',tags=,updated_at=2024-09-18T01:36:34Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1318.096682] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1318.096851] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1318.097051] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1318.097209] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1318.097363] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1318.097575] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1318.097774] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1318.097953] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1318.098140] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1318.098323] env[61985]: DEBUG nova.virt.hardware [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1318.099221] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b77678-b6e3-4ed6-b4c9-ff5188a8b644 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.107625] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf79e962-21e7-4ee1-937a-732d948d07c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.121711] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:8c:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e55c248-c504-4c7a-bbe9-f42cf417aee7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a71d390e-e1df-4d20-a37a-e5a2331f9b7f', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1318.129255] env[61985]: DEBUG oslo.service.loopingcall [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1318.129480] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1318.129686] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c8bdb0d-ef5e-4717-afbb-4be324143477 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.148412] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1318.148412] env[61985]: value = "task-936873" [ 1318.148412] env[61985]: _type = "Task" [ 1318.148412] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.155964] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936873, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.658505] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936873, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1318.801259] env[61985]: DEBUG nova.compute.manager [req-6d4a106c-f9aa-4f73-8ca7-c8afac8cc573 req-11b3fd02-54a2-4cca-b05b-316d2416b5e8 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received event network-changed-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1318.801501] env[61985]: DEBUG nova.compute.manager [req-6d4a106c-f9aa-4f73-8ca7-c8afac8cc573 req-11b3fd02-54a2-4cca-b05b-316d2416b5e8 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Refreshing instance network info cache due to event network-changed-a71d390e-e1df-4d20-a37a-e5a2331f9b7f. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1318.801719] env[61985]: DEBUG oslo_concurrency.lockutils [req-6d4a106c-f9aa-4f73-8ca7-c8afac8cc573 req-11b3fd02-54a2-4cca-b05b-316d2416b5e8 service nova] Acquiring lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1318.801870] env[61985]: DEBUG oslo_concurrency.lockutils [req-6d4a106c-f9aa-4f73-8ca7-c8afac8cc573 req-11b3fd02-54a2-4cca-b05b-316d2416b5e8 service nova] Acquired lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1318.802212] env[61985]: DEBUG nova.network.neutron [req-6d4a106c-f9aa-4f73-8ca7-c8afac8cc573 req-11b3fd02-54a2-4cca-b05b-316d2416b5e8 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Refreshing network info cache for port a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1319.160631] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936873, 'name': CreateVM_Task} progress is 99%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.561225] env[61985]: DEBUG nova.network.neutron [req-6d4a106c-f9aa-4f73-8ca7-c8afac8cc573 req-11b3fd02-54a2-4cca-b05b-316d2416b5e8 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updated VIF entry in instance network info cache for port a71d390e-e1df-4d20-a37a-e5a2331f9b7f. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1319.561642] env[61985]: DEBUG nova.network.neutron [req-6d4a106c-f9aa-4f73-8ca7-c8afac8cc573 req-11b3fd02-54a2-4cca-b05b-316d2416b5e8 service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating instance_info_cache with network_info: [{"id": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "address": "fa:16:3e:e1:8c:42", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71d390e-e1", "ovs_interfaceid": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1319.659571] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936873, 'name': CreateVM_Task, 'duration_secs': 1.285532} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.659741] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1319.660403] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1319.660577] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.660968] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1319.661227] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46ab483f-7b0a-4a7b-ab35-a85878b483b4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.665450] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1319.665450] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a911a9-d7d1-6340-d343-bd99392fb41d" [ 1319.665450] env[61985]: _type = "Task" [ 1319.665450] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.672420] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a911a9-d7d1-6340-d343-bd99392fb41d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.725341] env[61985]: INFO nova.compute.manager [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Rebuilding instance [ 1319.760349] env[61985]: DEBUG nova.compute.manager [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1319.760921] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1aa97e-bda7-4d5d-8cbc-cc33a4df63c3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.064215] env[61985]: DEBUG oslo_concurrency.lockutils [req-6d4a106c-f9aa-4f73-8ca7-c8afac8cc573 req-11b3fd02-54a2-4cca-b05b-316d2416b5e8 service nova] Releasing lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1320.177157] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1320.177501] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Processing image c4228803-7851-47f8-b83e-fc58a07898d8 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1320.177651] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8/c4228803-7851-47f8-b83e-fc58a07898d8.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1320.177822] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8/c4228803-7851-47f8-b83e-fc58a07898d8.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1320.178020] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1320.178274] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a66680c-7482-4b1b-9a02-531b0c4c7dd9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.195247] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1320.195477] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1320.196134] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ad61982-bfb0-40e5-87a4-8af1066cefdb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.201352] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1320.201352] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527783a2-e2dd-82d5-0830-cc446ad5e9e8" [ 1320.201352] env[61985]: _type = "Task" [ 1320.201352] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.208496] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]527783a2-e2dd-82d5-0830-cc446ad5e9e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.271647] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1320.271927] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-974b74d8-0c27-4ace-8138-82783fc0721c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.278230] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for the task: (returnval){ [ 1320.278230] env[61985]: value = "task-936874" [ 1320.278230] env[61985]: _type = "Task" [ 1320.278230] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.285718] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.712120] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Preparing fetch location {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1320.712396] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Fetch image to [datastore1] OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f/OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f.vmdk {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1320.712589] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Downloading stream optimized image c4228803-7851-47f8-b83e-fc58a07898d8 to [datastore1] OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f/OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f.vmdk on the data store datastore1 as vApp {{(pid=61985) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1320.712768] env[61985]: DEBUG nova.virt.vmwareapi.images [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Downloading image file data c4228803-7851-47f8-b83e-fc58a07898d8 to the ESX as VM named 'OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f' {{(pid=61985) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1320.782949] env[61985]: DEBUG oslo_vmware.rw_handles [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1320.782949] env[61985]: value = "resgroup-9" [ 1320.782949] env[61985]: _type = "ResourcePool" [ 1320.782949] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1320.783236] env[61985]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-68415b7a-8648-4988-a7c0-c057b81a7a6b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.802760] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936874, 'name': PowerOffVM_Task, 'duration_secs': 0.226948} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.803920] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1320.804588] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1320.804911] env[61985]: DEBUG oslo_vmware.rw_handles [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lease: (returnval){ [ 1320.804911] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52130c53-2140-4a4d-8d52-5b747e346060" [ 1320.804911] env[61985]: _type = "HttpNfcLease" [ 1320.804911] env[61985]: } obtained for vApp import into resource pool (val){ [ 1320.804911] env[61985]: value = "resgroup-9" [ 1320.804911] env[61985]: _type = "ResourcePool" [ 1320.804911] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1320.805229] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the lease: (returnval){ [ 1320.805229] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52130c53-2140-4a4d-8d52-5b747e346060" [ 1320.805229] env[61985]: _type = "HttpNfcLease" [ 1320.805229] env[61985]: } to be ready. {{(pid=61985) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1320.805367] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdbf99f7-2fe0-4baa-b906-92eddd2e498e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.813110] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1320.813110] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52130c53-2140-4a4d-8d52-5b747e346060" [ 1320.813110] env[61985]: _type = "HttpNfcLease" [ 1320.813110] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1320.814134] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for the task: (returnval){ [ 1320.814134] env[61985]: value = "task-936876" [ 1320.814134] env[61985]: _type = "Task" [ 1320.814134] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.821994] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936876, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.315025] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1321.315025] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52130c53-2140-4a4d-8d52-5b747e346060" [ 1321.315025] env[61985]: _type = "HttpNfcLease" [ 1321.315025] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1321.322836] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1321.323075] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1321.323325] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211518', 'volume_id': 'a1fed62e-6782-4ff4-b5de-21f1669862fc', 'name': 'volume-a1fed62e-6782-4ff4-b5de-21f1669862fc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd74cbed6-8f70-4e8d-b050-23e00ac334d4', 'attached_at': '', 'detached_at': '', 'volume_id': 'a1fed62e-6782-4ff4-b5de-21f1669862fc', 'serial': 'a1fed62e-6782-4ff4-b5de-21f1669862fc'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1321.324052] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783f9c4c-cb0a-4f2b-9cf7-7044fa1d6c31 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.341785] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748c1211-a905-4747-aa30-8efecedd337c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.347917] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbded56-a79e-42d5-87e4-7fa16378f338 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.365047] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ada1a2-5918-45af-a532-bb4ebc492c97 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.380999] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] The volume has not been displaced from its original location: [datastore1] volume-a1fed62e-6782-4ff4-b5de-21f1669862fc/volume-a1fed62e-6782-4ff4-b5de-21f1669862fc.vmdk. No consolidation needed. {{(pid=61985) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1321.386306] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Reconfiguring VM instance instance-00000075 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1321.386590] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e529d17f-6e61-4e04-8a1d-c78e3096845f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.403593] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for the task: (returnval){ [ 1321.403593] env[61985]: value = "task-936877" [ 1321.403593] env[61985]: _type = "Task" [ 1321.403593] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.414359] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936877, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.815398] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1321.815398] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52130c53-2140-4a4d-8d52-5b747e346060" [ 1321.815398] env[61985]: _type = "HttpNfcLease" [ 1321.815398] env[61985]: } is initializing. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1321.912647] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936877, 'name': ReconfigVM_Task, 'duration_secs': 0.215105} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.912919] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Reconfigured VM instance instance-00000075 to detach disk 2000 {{(pid=61985) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1321.917451] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a928516f-10a7-4d04-9ba1-2388122344cb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.932163] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for the task: (returnval){ [ 1321.932163] env[61985]: value = "task-936878" [ 1321.932163] env[61985]: _type = "Task" [ 1321.932163] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.939537] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936878, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.317186] env[61985]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1322.317186] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52130c53-2140-4a4d-8d52-5b747e346060" [ 1322.317186] env[61985]: _type = "HttpNfcLease" [ 1322.317186] env[61985]: } is ready. {{(pid=61985) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1322.317652] env[61985]: DEBUG oslo_vmware.rw_handles [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1322.317652] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52130c53-2140-4a4d-8d52-5b747e346060" [ 1322.317652] env[61985]: _type = "HttpNfcLease" [ 1322.317652] env[61985]: }. {{(pid=61985) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1322.318755] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a20d8e-09b0-4155-aeb6-bfa451b81631 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.327639] env[61985]: DEBUG oslo_vmware.rw_handles [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f81ef7-3edd-89d6-9700-99beafbf5d70/disk-0.vmdk from lease info. {{(pid=61985) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1322.327842] env[61985]: DEBUG oslo_vmware.rw_handles [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating HTTP connection to write to file with size = 31666688 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f81ef7-3edd-89d6-9700-99beafbf5d70/disk-0.vmdk. {{(pid=61985) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1322.390350] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-dd8fc9ee-62f4-4f08-a831-b985081bacd8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.440801] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936878, 'name': ReconfigVM_Task, 'duration_secs': 0.108369} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.441103] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-211518', 'volume_id': 'a1fed62e-6782-4ff4-b5de-21f1669862fc', 'name': 'volume-a1fed62e-6782-4ff4-b5de-21f1669862fc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd74cbed6-8f70-4e8d-b050-23e00ac334d4', 'attached_at': '', 'detached_at': '', 'volume_id': 'a1fed62e-6782-4ff4-b5de-21f1669862fc', 'serial': 'a1fed62e-6782-4ff4-b5de-21f1669862fc'} {{(pid=61985) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1322.441387] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1322.442137] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a00291-3d57-4b23-a616-b2290986c8f1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.448453] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1322.448679] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e134975a-11bd-4aca-8822-3d5c21d09cb3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.515141] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1322.515390] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1322.515573] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Deleting the datastore file [datastore1] d74cbed6-8f70-4e8d-b050-23e00ac334d4 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1322.515895] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40dd1ba1-f160-48f5-bcff-b0287b8d863f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.522072] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for the task: (returnval){ [ 1322.522072] env[61985]: value = "task-936880" [ 1322.522072] env[61985]: _type = "Task" [ 1322.522072] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.530361] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936880, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.034245] env[61985]: DEBUG oslo_vmware.api [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Task: {'id': task-936880, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.080141} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.036566] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1323.036799] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1323.037055] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1323.097250] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Volume detach. Driver type: vmdk {{(pid=61985) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1323.097611] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1017e655-19b3-4eac-a9df-843fa9837261 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.108959] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d93e62a-6338-4f9e-a8c3-c3fb0f8e2756 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.137651] env[61985]: ERROR nova.compute.manager [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Failed to detach volume a1fed62e-6782-4ff4-b5de-21f1669862fc from /dev/sda: nova.exception.InstanceNotFound: Instance d74cbed6-8f70-4e8d-b050-23e00ac334d4 could not be found. [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Traceback (most recent call last): [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 4148, in _do_rebuild_instance [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self.driver.rebuild(**kwargs) [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] raise NotImplementedError() [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] NotImplementedError [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] During handling of the above exception, another exception occurred: [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Traceback (most recent call last): [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 3571, in _detach_root_volume [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self.driver.detach_volume(context, old_connection_info, [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] return self._volumeops.detach_volume(connection_info, instance) [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self._detach_volume_vmdk(connection_info, instance) [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] stable_ref.fetch_moref(session) [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] nova.exception.InstanceNotFound: Instance d74cbed6-8f70-4e8d-b050-23e00ac334d4 could not be found. [ 1323.137651] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] [ 1323.187267] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.187494] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.187657] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1323.270694] env[61985]: DEBUG nova.compute.utils [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Build of instance d74cbed6-8f70-4e8d-b050-23e00ac334d4 aborted: Failed to rebuild volume backed instance. {{(pid=61985) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1323.273222] env[61985]: ERROR nova.compute.manager [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance d74cbed6-8f70-4e8d-b050-23e00ac334d4 aborted: Failed to rebuild volume backed instance. [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Traceback (most recent call last): [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 4148, in _do_rebuild_instance [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self.driver.rebuild(**kwargs) [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] raise NotImplementedError() [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] NotImplementedError [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] During handling of the above exception, another exception occurred: [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Traceback (most recent call last): [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 3606, in _rebuild_volume_backed_instance [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self._detach_root_volume(context, instance, root_bdm) [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 3585, in _detach_root_volume [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] with excutils.save_and_reraise_exception(): [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self.force_reraise() [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] raise self.value [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 3571, in _detach_root_volume [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self.driver.detach_volume(context, old_connection_info, [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] return self._volumeops.detach_volume(connection_info, instance) [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self._detach_volume_vmdk(connection_info, instance) [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] stable_ref.fetch_moref(session) [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] nova.exception.InstanceNotFound: Instance d74cbed6-8f70-4e8d-b050-23e00ac334d4 could not be found. [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] During handling of the above exception, another exception occurred: [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Traceback (most recent call last): [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 10869, in _error_out_instance_on_exception [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] yield [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 3874, in rebuild_instance [ 1323.273222] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self._do_rebuild_instance_with_claim( [ 1323.274595] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 3960, in _do_rebuild_instance_with_claim [ 1323.274595] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self._do_rebuild_instance( [ 1323.274595] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 4152, in _do_rebuild_instance [ 1323.274595] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self._rebuild_default_impl(**kwargs) [ 1323.274595] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 3729, in _rebuild_default_impl [ 1323.274595] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] self._rebuild_volume_backed_instance( [ 1323.274595] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] File "/opt/stack/nova/nova/compute/manager.py", line 3621, in _rebuild_volume_backed_instance [ 1323.274595] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] raise exception.BuildAbortException( [ 1323.274595] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] nova.exception.BuildAbortException: Build of instance d74cbed6-8f70-4e8d-b050-23e00ac334d4 aborted: Failed to rebuild volume backed instance. [ 1323.274595] env[61985]: ERROR nova.compute.manager [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] [ 1323.461464] env[61985]: DEBUG oslo_vmware.rw_handles [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Completed reading data from the image iterator. {{(pid=61985) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1323.461981] env[61985]: DEBUG oslo_vmware.rw_handles [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f81ef7-3edd-89d6-9700-99beafbf5d70/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1323.462948] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415880d5-73e8-4d8a-a8dc-1afb44433cd6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.469666] env[61985]: DEBUG oslo_vmware.rw_handles [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f81ef7-3edd-89d6-9700-99beafbf5d70/disk-0.vmdk is in state: ready. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1323.469931] env[61985]: DEBUG oslo_vmware.rw_handles [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f81ef7-3edd-89d6-9700-99beafbf5d70/disk-0.vmdk. {{(pid=61985) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1323.470244] env[61985]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ce5e9e0c-4656-487d-9f75-4e597bd2d237 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.666403] env[61985]: DEBUG oslo_vmware.rw_handles [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f81ef7-3edd-89d6-9700-99beafbf5d70/disk-0.vmdk. {{(pid=61985) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1323.666639] env[61985]: INFO nova.virt.vmwareapi.images [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Downloaded image file data c4228803-7851-47f8-b83e-fc58a07898d8 [ 1323.667505] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1cd7d7b-6765-40b1-8522-a3c85bdbf2b1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.682776] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3fb29422-3f11-4666-a72a-8f98e50a9d60 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.693714] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1323.693844] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquired lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1323.693993] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Forcefully refreshing network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1323.709998] env[61985]: INFO nova.virt.vmwareapi.images [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] The imported VM was unregistered [ 1323.712376] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Caching image {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1323.712577] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Creating directory with path [datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8 {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1323.713082] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3cba771-20db-437a-882b-8dd52576eaf6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.723211] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Created directory with path [datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8 {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1323.723405] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f/OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f.vmdk to [datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8/c4228803-7851-47f8-b83e-fc58a07898d8.vmdk. {{(pid=61985) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1323.723639] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-4b062800-fa81-48b4-8bc6-8380d29dcef1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.730505] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1323.730505] env[61985]: value = "task-936882" [ 1323.730505] env[61985]: _type = "Task" [ 1323.730505] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.738171] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936882, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.240784] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936882, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.741493] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936882, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.973875] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating instance_info_cache with network_info: [{"id": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "address": "fa:16:3e:e1:8c:42", "network": {"id": "3ee1dd7c-797b-4238-b7f7-731cb7bbed68", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-275903810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7e22f7d7b7b407ebb2d00e174d8a42c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71d390e-e1", "ovs_interfaceid": "a71d390e-e1df-4d20-a37a-e5a2331f9b7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1325.243703] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936882, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.287801] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.288098] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.306889] env[61985]: DEBUG nova.scheduler.client.report [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Refreshing inventories for resource provider aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1325.320540] env[61985]: DEBUG nova.scheduler.client.report [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Updating ProviderTree inventory for provider aed7e5b3-c662-4538-8447-c4f67b460215 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1325.320792] env[61985]: DEBUG nova.compute.provider_tree [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Updating inventory in ProviderTree for provider aed7e5b3-c662-4538-8447-c4f67b460215 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1325.334036] env[61985]: DEBUG nova.scheduler.client.report [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Refreshing aggregate associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, aggregates: None {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1325.353519] env[61985]: DEBUG nova.scheduler.client.report [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Refreshing trait associations for resource provider aed7e5b3-c662-4538-8447-c4f67b460215, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61985) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1325.397173] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92cf968-d56f-4974-935e-1f25c9f4d0e5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.405468] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b020ac-12fc-47ee-b4f8-1c03e0a273f5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.435551] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f98c35-7029-4b38-aedb-3ae777d52e34 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.443523] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd55c442-ce16-4072-bb4e-19f54db2b80a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.457306] env[61985]: DEBUG nova.compute.provider_tree [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1325.476392] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Releasing lock "refresh_cache-8c95afa2-dbec-4408-a27d-595431e09748" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1325.476607] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updated the network info_cache for instance {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10003}} [ 1325.476813] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.476980] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.477160] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.477309] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.477440] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 1325.477583] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.743465] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936882, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.755618] env[61985]: DEBUG oslo_concurrency.lockutils [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Acquiring lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.756037] env[61985]: DEBUG oslo_concurrency.lockutils [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.756282] env[61985]: DEBUG oslo_concurrency.lockutils [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Acquiring lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.756481] env[61985]: DEBUG oslo_concurrency.lockutils [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.756698] env[61985]: DEBUG oslo_concurrency.lockutils [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1325.759134] env[61985]: INFO nova.compute.manager [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Terminating instance [ 1325.761357] env[61985]: DEBUG nova.compute.manager [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1325.761659] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21cbfbfc-87e0-4cc5-9f2e-6f6bc1b196f8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.770852] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2b2cea-bc4b-4a46-945b-b714a2984d15 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.796233] env[61985]: WARNING nova.virt.vmwareapi.driver [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance d74cbed6-8f70-4e8d-b050-23e00ac334d4 could not be found. [ 1325.796468] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1325.796807] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e9438e5-47ab-4264-b5b1-ba2ae94023bb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.804758] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0629b2-35c1-427e-b306-d094951c5902 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.828160] env[61985]: WARNING nova.virt.vmwareapi.vmops [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d74cbed6-8f70-4e8d-b050-23e00ac334d4 could not be found. [ 1325.828390] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1325.828587] env[61985]: INFO nova.compute.manager [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Took 0.07 seconds to destroy the instance on the hypervisor. [ 1325.828864] env[61985]: DEBUG oslo.service.loopingcall [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1325.829126] env[61985]: DEBUG nova.compute.manager [-] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1325.829228] env[61985]: DEBUG nova.network.neutron [-] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1325.961008] env[61985]: DEBUG nova.scheduler.client.report [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1325.982719] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.243863] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936882, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.394987} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1326.244236] env[61985]: INFO nova.virt.vmwareapi.ds_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f/OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f.vmdk to [datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8/c4228803-7851-47f8-b83e-fc58a07898d8.vmdk. [ 1326.244445] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Cleaning up location [datastore1] OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1326.244616] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_d9297d88-5bf8-4af5-9e1a-b6b84d7d356f {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1326.244869] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a3abae4a-66bb-465f-bd4b-c211bc083488 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.250965] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1326.250965] env[61985]: value = "task-936883" [ 1326.250965] env[61985]: _type = "Task" [ 1326.250965] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.258055] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936883, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.314693] env[61985]: DEBUG nova.compute.manager [req-9abd6b08-d79f-4e89-983b-a5e2879131cf req-15e0c9d9-a44a-4dad-81bb-082f590851c9 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Received event network-vif-deleted-8138f233-84b3-454c-a61a-4de873fa461b {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1326.314893] env[61985]: INFO nova.compute.manager [req-9abd6b08-d79f-4e89-983b-a5e2879131cf req-15e0c9d9-a44a-4dad-81bb-082f590851c9 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Neutron deleted interface 8138f233-84b3-454c-a61a-4de873fa461b; detaching it from the instance and deleting it from the info cache [ 1326.315049] env[61985]: DEBUG nova.network.neutron [req-9abd6b08-d79f-4e89-983b-a5e2879131cf req-15e0c9d9-a44a-4dad-81bb-082f590851c9 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1326.465887] env[61985]: DEBUG oslo_concurrency.lockutils [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.178s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.466833] env[61985]: INFO nova.compute.manager [None req-ee4f3686-cd63-40e5-9a84-fdd4622e458b tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Successfully reverted task state from rebuilding on failure for instance. [ 1326.470920] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.490s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1326.471115] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1326.471264] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1326.472720] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288d3341-9f5e-43cb-90b1-55dc3487ac4a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.484446] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc6773f-f3bb-4081-b510-18799aa9640a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.500031] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58ec2b8-4823-4b3a-9606-806a8627b7ae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.506721] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0932a0-89d5-47bc-9934-cf4b1ea98f34 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.537168] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181161MB free_disk=43GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1326.537345] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1326.537551] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1326.761552] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936883, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.045776} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1326.762832] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1326.762832] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8/c4228803-7851-47f8-b83e-fc58a07898d8.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1326.762832] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8/c4228803-7851-47f8-b83e-fc58a07898d8.vmdk to [datastore1] 8c95afa2-dbec-4408-a27d-595431e09748/8c95afa2-dbec-4408-a27d-595431e09748.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1326.763038] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9530066-3713-4845-bfbb-1ab423ea53bf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.769559] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1326.769559] env[61985]: value = "task-936884" [ 1326.769559] env[61985]: _type = "Task" [ 1326.769559] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.777025] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.794482] env[61985]: DEBUG nova.network.neutron [-] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1326.818217] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44fd3f3f-d2fa-4d9f-b2b6-5abf1ab04ac6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.828215] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac159d82-bd42-4225-8295-1ea8ff4e1278 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.856336] env[61985]: DEBUG nova.compute.manager [req-9abd6b08-d79f-4e89-983b-a5e2879131cf req-15e0c9d9-a44a-4dad-81bb-082f590851c9 service nova] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Detach interface failed, port_id=8138f233-84b3-454c-a61a-4de873fa461b, reason: Instance d74cbed6-8f70-4e8d-b050-23e00ac334d4 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1327.280525] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936884, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.297328] env[61985]: INFO nova.compute.manager [-] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Took 1.47 seconds to deallocate network for instance. [ 1327.567028] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 4908dca0-b19f-4fff-b560-075ad46020d3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1327.567028] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 8c95afa2-dbec-4408-a27d-595431e09748 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1327.567216] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1327.567338] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1327.609163] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b088789-32c1-4dc6-8495-2f44388edfa6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.616896] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b614fc70-d9cd-4813-a481-1dc21632c09e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.648812] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a86fb4-d4b7-4ad2-9bb7-af8c066bb4de {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.656560] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86514ccf-97e4-4be1-b675-cc180763b035 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.671016] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1327.780343] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936884, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.851150] env[61985]: INFO nova.compute.manager [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Took 0.55 seconds to detach 1 volumes for instance. [ 1327.851901] env[61985]: DEBUG nova.compute.manager [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] [instance: d74cbed6-8f70-4e8d-b050-23e00ac334d4] Deleting volume: a1fed62e-6782-4ff4-b5de-21f1669862fc {{(pid=61985) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3253}} [ 1328.174686] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1328.281739] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936884, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.400383] env[61985]: DEBUG oslo_concurrency.lockutils [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1328.681040] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1328.681352] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.144s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1328.681735] env[61985]: DEBUG oslo_concurrency.lockutils [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.282s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1328.681988] env[61985]: DEBUG nova.objects.instance [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lazy-loading 'resources' on Instance uuid d74cbed6-8f70-4e8d-b050-23e00ac334d4 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1328.781699] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936884, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.225693] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119662b3-0578-4445-9f52-5af0e03a6448 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.232890] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650d692c-ae06-4b02-9dc9-fb6bbaf11cae {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.263832] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9911cc4d-de51-497d-b9ea-294c219211bd {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.271139] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95301d0-d097-4173-abd9-9d94d50850a3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.286625] env[61985]: DEBUG nova.compute.provider_tree [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1329.290764] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936884, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.29916} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.291023] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c4228803-7851-47f8-b83e-fc58a07898d8/c4228803-7851-47f8-b83e-fc58a07898d8.vmdk to [datastore1] 8c95afa2-dbec-4408-a27d-595431e09748/8c95afa2-dbec-4408-a27d-595431e09748.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1329.291819] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa083d68-0e79-4e5e-b10b-8f94cf8fb652 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.313010] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 8c95afa2-dbec-4408-a27d-595431e09748/8c95afa2-dbec-4408-a27d-595431e09748.vmdk or device None with type streamOptimized {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1329.313845] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37d9ad89-67a8-48f5-9a45-df630efbbcd2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.332971] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1329.332971] env[61985]: value = "task-936886" [ 1329.332971] env[61985]: _type = "Task" [ 1329.332971] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.344665] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936886, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.697220] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.697452] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.795861] env[61985]: DEBUG nova.scheduler.client.report [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1329.843593] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936886, 'name': ReconfigVM_Task, 'duration_secs': 0.266398} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.843902] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 8c95afa2-dbec-4408-a27d-595431e09748/8c95afa2-dbec-4408-a27d-595431e09748.vmdk or device None with type streamOptimized {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1329.844596] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e9c6b982-d2a4-4aeb-b095-ef11c636ada8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.851326] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1329.851326] env[61985]: value = "task-936887" [ 1329.851326] env[61985]: _type = "Task" [ 1329.851326] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1329.858670] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936887, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.301101] env[61985]: DEBUG oslo_concurrency.lockutils [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.363571] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936887, 'name': Rename_Task, 'duration_secs': 0.137087} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.363571] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1330.363790] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5fd17928-2186-4674-ad75-7e7c5c1b69d5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.370067] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1330.370067] env[61985]: value = "task-936888" [ 1330.370067] env[61985]: _type = "Task" [ 1330.370067] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.377146] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936888, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.819134] env[61985]: DEBUG oslo_concurrency.lockutils [None req-937e3ead-c488-417a-8f2e-b36ccfac795a tempest-ServerActionsV293TestJSON-1583810653 tempest-ServerActionsV293TestJSON-1583810653-project-member] Lock "d74cbed6-8f70-4e8d-b050-23e00ac334d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.063s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.881016] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936888, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.380180] env[61985]: DEBUG oslo_vmware.api [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936888, 'name': PowerOnVM_Task, 'duration_secs': 0.514766} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.380454] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1331.469765] env[61985]: DEBUG nova.compute.manager [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1331.470750] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca9c001-b92b-4165-a5cd-dddd239b1c37 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.986660] env[61985]: DEBUG oslo_concurrency.lockutils [None req-bee654d6-eed3-445e-be19-bc066eecb45c tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "8c95afa2-dbec-4408-a27d-595431e09748" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.292s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.319714] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "4908dca0-b19f-4fff-b560-075ad46020d3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1346.320133] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "4908dca0-b19f-4fff-b560-075ad46020d3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1346.320193] env[61985]: DEBUG nova.compute.manager [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1346.321112] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca14d834-4018-483f-823d-0d687d7990ff {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.327822] env[61985]: DEBUG nova.compute.manager [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61985) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3374}} [ 1346.328378] env[61985]: DEBUG nova.objects.instance [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'flavor' on Instance uuid 4908dca0-b19f-4fff-b560-075ad46020d3 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1346.834611] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1346.834903] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c3bc0035-7ac5-4458-afd8-3292f1ad394c {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.842533] env[61985]: DEBUG oslo_vmware.api [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1346.842533] env[61985]: value = "task-936889" [ 1346.842533] env[61985]: _type = "Task" [ 1346.842533] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.850591] env[61985]: DEBUG oslo_vmware.api [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.352763] env[61985]: DEBUG oslo_vmware.api [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936889, 'name': PowerOffVM_Task, 'duration_secs': 0.158637} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.353119] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1347.353214] env[61985]: DEBUG nova.compute.manager [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1347.353952] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac248260-5f2e-4fe5-9603-36825c39ae36 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.865572] env[61985]: DEBUG oslo_concurrency.lockutils [None req-fe07fa96-b566-4e27-a63d-779d3baa07ea tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "4908dca0-b19f-4fff-b560-075ad46020d3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1348.719645] env[61985]: DEBUG nova.objects.instance [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'flavor' on Instance uuid 4908dca0-b19f-4fff-b560-075ad46020d3 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1349.228071] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1349.228071] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1349.228071] env[61985]: DEBUG nova.network.neutron [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1349.228071] env[61985]: DEBUG nova.objects.instance [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'info_cache' on Instance uuid 4908dca0-b19f-4fff-b560-075ad46020d3 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1349.732230] env[61985]: DEBUG nova.objects.base [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Object Instance<4908dca0-b19f-4fff-b560-075ad46020d3> lazy-loaded attributes: flavor,info_cache {{(pid=61985) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1350.438893] env[61985]: DEBUG nova.network.neutron [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Updating instance_info_cache with network_info: [{"id": "1311d04d-d434-4c2a-ac57-ce6c00626037", "address": "fa:16:3e:db:23:37", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1311d04d-d4", "ovs_interfaceid": "1311d04d-d434-4c2a-ac57-ce6c00626037", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1350.942906] env[61985]: DEBUG oslo_concurrency.lockutils [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1351.445822] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1351.446181] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-647082b1-855d-4336-8b47-ada69e6d7a62 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.453723] env[61985]: DEBUG oslo_vmware.api [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1351.453723] env[61985]: value = "task-936890" [ 1351.453723] env[61985]: _type = "Task" [ 1351.453723] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1351.461088] env[61985]: DEBUG oslo_vmware.api [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936890, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1351.964126] env[61985]: DEBUG oslo_vmware.api [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936890, 'name': PowerOnVM_Task, 'duration_secs': 0.392328} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1351.964435] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1351.964625] env[61985]: DEBUG nova.compute.manager [None req-d4eab90e-32e7-429b-b1bb-48b5ace65296 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1351.965409] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eeea160-0ff5-4411-8170-42e1f3250173 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.692608] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233547b1-86be-40ac-a3fa-3aabb0a87a7a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.699593] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f017b8-0308-4851-843e-904f830a1f2c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Suspending the VM {{(pid=61985) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1353.699836] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-720159d7-5e4c-4211-b3ea-4793b842eb41 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.705339] env[61985]: DEBUG oslo_vmware.api [None req-e6f017b8-0308-4851-843e-904f830a1f2c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1353.705339] env[61985]: value = "task-936891" [ 1353.705339] env[61985]: _type = "Task" [ 1353.705339] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1353.712904] env[61985]: DEBUG oslo_vmware.api [None req-e6f017b8-0308-4851-843e-904f830a1f2c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936891, 'name': SuspendVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.215890] env[61985]: DEBUG oslo_vmware.api [None req-e6f017b8-0308-4851-843e-904f830a1f2c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936891, 'name': SuspendVM_Task} progress is 70%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.716519] env[61985]: DEBUG oslo_vmware.api [None req-e6f017b8-0308-4851-843e-904f830a1f2c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936891, 'name': SuspendVM_Task, 'duration_secs': 0.578803} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.716964] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f017b8-0308-4851-843e-904f830a1f2c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Suspended the VM {{(pid=61985) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1354.717069] env[61985]: DEBUG nova.compute.manager [None req-e6f017b8-0308-4851-843e-904f830a1f2c tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1354.717889] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910207d9-f57e-43c6-90f1-e8db5c1917db {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.034825] env[61985]: INFO nova.compute.manager [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Resuming [ 1356.035602] env[61985]: DEBUG nova.objects.instance [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'flavor' on Instance uuid 4908dca0-b19f-4fff-b560-075ad46020d3 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1357.044760] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1357.044760] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquired lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.045015] env[61985]: DEBUG nova.network.neutron [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1357.742306] env[61985]: DEBUG nova.network.neutron [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Updating instance_info_cache with network_info: [{"id": "1311d04d-d434-4c2a-ac57-ce6c00626037", "address": "fa:16:3e:db:23:37", "network": {"id": "20c6ba62-2fbb-4b6c-9cf4-66c704f90635", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1125560142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "23a4809e353a4ecaa1bd53612fb4afc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1311d04d-d4", "ovs_interfaceid": "1311d04d-d434-4c2a-ac57-ce6c00626037", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1358.244833] env[61985]: DEBUG oslo_concurrency.lockutils [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Releasing lock "refresh_cache-4908dca0-b19f-4fff-b560-075ad46020d3" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1358.245876] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7d424e-1bc1-48e5-ae19-26af6e26c54f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.252757] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Resuming the VM {{(pid=61985) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1358.252979] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96d4a5d3-1152-40fe-b58a-df63a308c8c6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.259215] env[61985]: DEBUG oslo_vmware.api [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1358.259215] env[61985]: value = "task-936892" [ 1358.259215] env[61985]: _type = "Task" [ 1358.259215] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1358.266655] env[61985]: DEBUG oslo_vmware.api [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936892, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.769078] env[61985]: DEBUG oslo_vmware.api [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936892, 'name': PowerOnVM_Task, 'duration_secs': 0.470448} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1358.769389] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Resumed the VM {{(pid=61985) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1358.769593] env[61985]: DEBUG nova.compute.manager [None req-17180bba-32b5-4458-a937-088c3b6d1f73 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1358.770375] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebf65ab-8a9a-4497-9b7c-ce2e8abac946 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.631041] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "4908dca0-b19f-4fff-b560-075ad46020d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1359.631460] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "4908dca0-b19f-4fff-b560-075ad46020d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1359.631614] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "4908dca0-b19f-4fff-b560-075ad46020d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1359.631806] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "4908dca0-b19f-4fff-b560-075ad46020d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1359.631983] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "4908dca0-b19f-4fff-b560-075ad46020d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1359.634188] env[61985]: INFO nova.compute.manager [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Terminating instance [ 1359.635943] env[61985]: DEBUG nova.compute.manager [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1359.636170] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1359.637008] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718ba489-ed5c-45b4-8f93-c0ec9612a8b8 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.644771] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1359.645024] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-22e96a61-9049-4b4e-81cb-e70da79005d5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.651083] env[61985]: DEBUG oslo_vmware.api [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1359.651083] env[61985]: value = "task-936893" [ 1359.651083] env[61985]: _type = "Task" [ 1359.651083] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.658366] env[61985]: DEBUG oslo_vmware.api [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.161218] env[61985]: DEBUG oslo_vmware.api [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936893, 'name': PowerOffVM_Task, 'duration_secs': 0.173405} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.161484] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1360.161657] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1360.161899] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16c1aede-ae96-422a-98bd-52ea861a4768 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.222878] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1360.223152] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1360.223353] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleting the datastore file [datastore1] 4908dca0-b19f-4fff-b560-075ad46020d3 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1360.223624] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cd5b2de7-189c-4ad6-8a8f-a27eff2dba12 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.229752] env[61985]: DEBUG oslo_vmware.api [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for the task: (returnval){ [ 1360.229752] env[61985]: value = "task-936895" [ 1360.229752] env[61985]: _type = "Task" [ 1360.229752] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.237198] env[61985]: DEBUG oslo_vmware.api [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.739439] env[61985]: DEBUG oslo_vmware.api [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Task: {'id': task-936895, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138324} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.739834] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1360.739888] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1360.740072] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1360.740261] env[61985]: INFO nova.compute.manager [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1360.740516] env[61985]: DEBUG oslo.service.loopingcall [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1360.740711] env[61985]: DEBUG nova.compute.manager [-] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1360.740807] env[61985]: DEBUG nova.network.neutron [-] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1361.147794] env[61985]: DEBUG nova.compute.manager [req-632cdc10-6c9d-4950-900b-d6e636aba903 req-bef0c77d-3f76-47ef-9d64-687f5c56e52b service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Received event network-vif-deleted-1311d04d-d434-4c2a-ac57-ce6c00626037 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1361.148045] env[61985]: INFO nova.compute.manager [req-632cdc10-6c9d-4950-900b-d6e636aba903 req-bef0c77d-3f76-47ef-9d64-687f5c56e52b service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Neutron deleted interface 1311d04d-d434-4c2a-ac57-ce6c00626037; detaching it from the instance and deleting it from the info cache [ 1361.148231] env[61985]: DEBUG nova.network.neutron [req-632cdc10-6c9d-4950-900b-d6e636aba903 req-bef0c77d-3f76-47ef-9d64-687f5c56e52b service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1361.628314] env[61985]: DEBUG nova.network.neutron [-] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1361.650799] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2340889e-b853-438b-8b94-729c4bb054dc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.660007] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663e28dd-9d28-4db4-9b69-2e11f9fdc47d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.683209] env[61985]: DEBUG nova.compute.manager [req-632cdc10-6c9d-4950-900b-d6e636aba903 req-bef0c77d-3f76-47ef-9d64-687f5c56e52b service nova] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Detach interface failed, port_id=1311d04d-d434-4c2a-ac57-ce6c00626037, reason: Instance 4908dca0-b19f-4fff-b560-075ad46020d3 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1362.131665] env[61985]: INFO nova.compute.manager [-] [instance: 4908dca0-b19f-4fff-b560-075ad46020d3] Took 1.39 seconds to deallocate network for instance. [ 1362.638663] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1362.638912] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1362.639156] env[61985]: DEBUG nova.objects.instance [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lazy-loading 'resources' on Instance uuid 4908dca0-b19f-4fff-b560-075ad46020d3 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1363.184010] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74b0ea5-d8b4-494c-966d-2439da96956b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.191831] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5ee3be-c0e9-4b33-921e-d972c7b2e875 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.222827] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4152325c-3d41-4fc6-b39f-345fe6602384 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.229396] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d6bfd1-e447-48d4-9b1b-13690512e12b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.242266] env[61985]: DEBUG nova.compute.provider_tree [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1363.745677] env[61985]: DEBUG nova.scheduler.client.report [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1364.250952] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.612s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1364.269724] env[61985]: INFO nova.scheduler.client.report [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Deleted allocations for instance 4908dca0-b19f-4fff-b560-075ad46020d3 [ 1364.776988] env[61985]: DEBUG oslo_concurrency.lockutils [None req-8ebe5af6-5ebd-47bc-9fee-171f2b580423 tempest-ServerActionsTestJSON-1615869262 tempest-ServerActionsTestJSON-1615869262-project-member] Lock "4908dca0-b19f-4fff-b560-075ad46020d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.145s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1366.915886] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "8c95afa2-dbec-4408-a27d-595431e09748" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1366.916278] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "8c95afa2-dbec-4408-a27d-595431e09748" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1366.916608] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "8c95afa2-dbec-4408-a27d-595431e09748-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1366.916896] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "8c95afa2-dbec-4408-a27d-595431e09748-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1366.917169] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "8c95afa2-dbec-4408-a27d-595431e09748-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1366.919716] env[61985]: INFO nova.compute.manager [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Terminating instance [ 1366.921686] env[61985]: DEBUG nova.compute.manager [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1366.921775] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1366.922614] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a6e9d7-9b9e-41f7-a5df-82ce87195ba7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.930096] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1366.930486] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76351181-4236-45e3-8e7f-4afa86f1431a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.937219] env[61985]: DEBUG oslo_vmware.api [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1366.937219] env[61985]: value = "task-936896" [ 1366.937219] env[61985]: _type = "Task" [ 1366.937219] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.945205] env[61985]: DEBUG oslo_vmware.api [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936896, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.448197] env[61985]: DEBUG oslo_vmware.api [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936896, 'name': PowerOffVM_Task, 'duration_secs': 0.209297} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.448474] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1367.448651] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1367.448927] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8eb3ad55-200d-4f59-96dd-7ce74c2019f3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.506836] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1367.507039] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Deleting contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1367.507243] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleting the datastore file [datastore1] 8c95afa2-dbec-4408-a27d-595431e09748 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1367.507526] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89459e47-1ea5-485b-9b36-03050fef1414 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.514584] env[61985]: DEBUG oslo_vmware.api [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for the task: (returnval){ [ 1367.514584] env[61985]: value = "task-936898" [ 1367.514584] env[61985]: _type = "Task" [ 1367.514584] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.522132] env[61985]: DEBUG oslo_vmware.api [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.026941] env[61985]: DEBUG oslo_vmware.api [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Task: {'id': task-936898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196783} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1368.027586] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1368.027797] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Deleted contents of the VM from datastore datastore1 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1368.027977] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1368.028176] env[61985]: INFO nova.compute.manager [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1368.028431] env[61985]: DEBUG oslo.service.loopingcall [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1368.028632] env[61985]: DEBUG nova.compute.manager [-] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1368.028727] env[61985]: DEBUG nova.network.neutron [-] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1368.450847] env[61985]: DEBUG nova.compute.manager [req-dc3d5f5b-0641-42c5-869c-8c13090c5a83 req-798296c5-1944-4904-9f46-abbfb409d15d service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Received event network-vif-deleted-a71d390e-e1df-4d20-a37a-e5a2331f9b7f {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1368.450847] env[61985]: INFO nova.compute.manager [req-dc3d5f5b-0641-42c5-869c-8c13090c5a83 req-798296c5-1944-4904-9f46-abbfb409d15d service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Neutron deleted interface a71d390e-e1df-4d20-a37a-e5a2331f9b7f; detaching it from the instance and deleting it from the info cache [ 1368.450847] env[61985]: DEBUG nova.network.neutron [req-dc3d5f5b-0641-42c5-869c-8c13090c5a83 req-798296c5-1944-4904-9f46-abbfb409d15d service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1368.922769] env[61985]: DEBUG nova.network.neutron [-] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1368.953758] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64f67615-8881-4feb-b2c7-9eb2fe5a7747 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.964083] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f746a8a6-6693-4bb0-9eec-3a8df0f037d5 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.987159] env[61985]: DEBUG nova.compute.manager [req-dc3d5f5b-0641-42c5-869c-8c13090c5a83 req-798296c5-1944-4904-9f46-abbfb409d15d service nova] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Detach interface failed, port_id=a71d390e-e1df-4d20-a37a-e5a2331f9b7f, reason: Instance 8c95afa2-dbec-4408-a27d-595431e09748 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1369.426049] env[61985]: INFO nova.compute.manager [-] [instance: 8c95afa2-dbec-4408-a27d-595431e09748] Took 1.40 seconds to deallocate network for instance. [ 1369.476740] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquiring lock "752346f6-9093-4ed3-9722-94fa0ccdda98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1369.477020] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lock "752346f6-9093-4ed3-9722-94fa0ccdda98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1369.934263] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1369.934670] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1369.934913] env[61985]: DEBUG nova.objects.instance [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lazy-loading 'resources' on Instance uuid 8c95afa2-dbec-4408-a27d-595431e09748 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1369.979432] env[61985]: DEBUG nova.compute.manager [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Starting instance... {{(pid=61985) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1370.474778] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c2cd2c-0ca9-4ac7-ac4e-52692759241f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.482209] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a98cac4-7ac3-4ffb-9d6d-5d4a4e78ae5e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.517270] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1370.518058] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb560866-781b-4fab-833e-57859b0f789a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.524973] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e3b8a1-1b89-4a5a-a0b4-16cd1f9b15fb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.537711] env[61985]: DEBUG nova.compute.provider_tree [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1371.041052] env[61985]: DEBUG nova.scheduler.client.report [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1371.545690] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1371.548070] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.031s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1371.549849] env[61985]: INFO nova.compute.claims [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1371.563814] env[61985]: INFO nova.scheduler.client.report [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Deleted allocations for instance 8c95afa2-dbec-4408-a27d-595431e09748 [ 1372.071741] env[61985]: DEBUG oslo_concurrency.lockutils [None req-4c7b3200-a2a4-47e7-8aef-679dc6070628 tempest-AttachVolumeShelveTestJSON-996523539 tempest-AttachVolumeShelveTestJSON-996523539-project-member] Lock "8c95afa2-dbec-4408-a27d-595431e09748" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.155s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.589139] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9017975-bb74-4732-9803-0d5b879a0fcc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.597328] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1943f7aa-14e9-4033-92d3-ad15b7816ec2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.628907] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaea8515-3620-4771-a264-b65b7839f1f6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.635564] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ee4d07-0e39-4452-950e-e0f294897cbb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.648190] env[61985]: DEBUG nova.compute.provider_tree [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1373.151418] env[61985]: DEBUG nova.scheduler.client.report [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1373.656578] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.108s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1373.656955] env[61985]: DEBUG nova.compute.manager [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Start building networks asynchronously for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1374.163045] env[61985]: DEBUG nova.compute.utils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Using /dev/sd instead of None {{(pid=61985) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1374.164609] env[61985]: DEBUG nova.compute.manager [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Allocating IP information in the background. {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1374.164795] env[61985]: DEBUG nova.network.neutron [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] allocate_for_instance() {{(pid=61985) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1374.217483] env[61985]: DEBUG nova.policy [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb850d8f05884eadbad541ebcb26ec0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f73280838cc84d8b91b13a96d51fad53', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61985) authorize /opt/stack/nova/nova/policy.py:201}} [ 1374.455972] env[61985]: DEBUG nova.network.neutron [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Successfully created port: b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1374.668782] env[61985]: DEBUG nova.compute.manager [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Start building block device mappings for instance. {{(pid=61985) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1375.680997] env[61985]: DEBUG nova.compute.manager [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Start spawning the instance on the hypervisor. {{(pid=61985) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1375.709022] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-18T01:22:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-18T01:22:38Z,direct_url=,disk_format='vmdk',id=fe6d817b-6194-440f-988a-f4a94c580922,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='5c359de6550041d28450fc8d15586c31',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-18T01:22:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1375.709336] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Flavor limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1375.709495] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Image limits 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1375.709688] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Flavor pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1375.709843] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Image pref 0:0:0 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1375.709999] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61985) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1375.710277] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1375.710453] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1375.710628] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Got 1 possible topologies {{(pid=61985) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1375.710790] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1375.710976] env[61985]: DEBUG nova.virt.hardware [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61985) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1375.711894] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736be247-853a-404b-8ec3-c318ced6d607 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.720033] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9870272-859c-45a2-8b8c-043c6d532a16 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.849824] env[61985]: DEBUG nova.compute.manager [req-b2364480-495b-4fe7-9943-c740bcb06929 req-4b8715f5-8938-4dd9-b8f9-e28c165a279e service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Received event network-vif-plugged-b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1375.850304] env[61985]: DEBUG oslo_concurrency.lockutils [req-b2364480-495b-4fe7-9943-c740bcb06929 req-4b8715f5-8938-4dd9-b8f9-e28c165a279e service nova] Acquiring lock "752346f6-9093-4ed3-9722-94fa0ccdda98-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1375.850395] env[61985]: DEBUG oslo_concurrency.lockutils [req-b2364480-495b-4fe7-9943-c740bcb06929 req-4b8715f5-8938-4dd9-b8f9-e28c165a279e service nova] Lock "752346f6-9093-4ed3-9722-94fa0ccdda98-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1375.850516] env[61985]: DEBUG oslo_concurrency.lockutils [req-b2364480-495b-4fe7-9943-c740bcb06929 req-4b8715f5-8938-4dd9-b8f9-e28c165a279e service nova] Lock "752346f6-9093-4ed3-9722-94fa0ccdda98-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1375.850676] env[61985]: DEBUG nova.compute.manager [req-b2364480-495b-4fe7-9943-c740bcb06929 req-4b8715f5-8938-4dd9-b8f9-e28c165a279e service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] No waiting events found dispatching network-vif-plugged-b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1375.850875] env[61985]: WARNING nova.compute.manager [req-b2364480-495b-4fe7-9943-c740bcb06929 req-4b8715f5-8938-4dd9-b8f9-e28c165a279e service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Received unexpected event network-vif-plugged-b1e4bff2-85a4-4a99-aa99-072688a294b5 for instance with vm_state building and task_state spawning. [ 1375.939188] env[61985]: DEBUG nova.network.neutron [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Successfully updated port: b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1376.442376] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquiring lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1376.442586] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquired lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1376.442806] env[61985]: DEBUG nova.network.neutron [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1376.985871] env[61985]: DEBUG nova.network.neutron [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Instance cache missing network info. {{(pid=61985) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1377.187829] env[61985]: DEBUG nova.network.neutron [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updating instance_info_cache with network_info: [{"id": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "address": "fa:16:3e:67:e8:13", "network": {"id": "e5781e12-b404-4c77-ab10-4eacacdb6121", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-210244505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f73280838cc84d8b91b13a96d51fad53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e4bff2-85", "ovs_interfaceid": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1377.690600] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Releasing lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1377.691337] env[61985]: DEBUG nova.compute.manager [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Instance network_info: |[{"id": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "address": "fa:16:3e:67:e8:13", "network": {"id": "e5781e12-b404-4c77-ab10-4eacacdb6121", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-210244505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f73280838cc84d8b91b13a96d51fad53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e4bff2-85", "ovs_interfaceid": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61985) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1377.691841] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:e8:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b9aabc7c-0f6c-42eb-bd27-493a1496c0c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1e4bff2-85a4-4a99-aa99-072688a294b5', 'vif_model': 'vmxnet3'}] {{(pid=61985) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1377.700014] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Creating folder: Project (f73280838cc84d8b91b13a96d51fad53). Parent ref: group-v211285. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1377.700330] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4ad3364-663f-46fd-b4f5-88b0bb695741 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.711229] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Created folder: Project (f73280838cc84d8b91b13a96d51fad53) in parent group-v211285. [ 1377.711491] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Creating folder: Instances. Parent ref: group-v211526. {{(pid=61985) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1377.711766] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10a962b7-cd0e-491f-9b75-aacb3e872a03 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.720892] env[61985]: INFO nova.virt.vmwareapi.vm_util [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Created folder: Instances in parent group-v211526. [ 1377.721109] env[61985]: DEBUG oslo.service.loopingcall [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1377.721299] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Creating VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1377.721494] env[61985]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-650eead8-2c97-4a84-b998-e7c5c6ab47eb {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.741943] env[61985]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1377.741943] env[61985]: value = "task-936902" [ 1377.741943] env[61985]: _type = "Task" [ 1377.741943] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1377.749671] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936902, 'name': CreateVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1377.882768] env[61985]: DEBUG nova.compute.manager [req-bf5edcde-5ebe-4911-ab60-29b4e6d6d02b req-f65c40fb-a89b-4ad2-8363-6c66bfbff7ec service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Received event network-changed-b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1377.883022] env[61985]: DEBUG nova.compute.manager [req-bf5edcde-5ebe-4911-ab60-29b4e6d6d02b req-f65c40fb-a89b-4ad2-8363-6c66bfbff7ec service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Refreshing instance network info cache due to event network-changed-b1e4bff2-85a4-4a99-aa99-072688a294b5. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1377.883273] env[61985]: DEBUG oslo_concurrency.lockutils [req-bf5edcde-5ebe-4911-ab60-29b4e6d6d02b req-f65c40fb-a89b-4ad2-8363-6c66bfbff7ec service nova] Acquiring lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1377.883420] env[61985]: DEBUG oslo_concurrency.lockutils [req-bf5edcde-5ebe-4911-ab60-29b4e6d6d02b req-f65c40fb-a89b-4ad2-8363-6c66bfbff7ec service nova] Acquired lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1377.883616] env[61985]: DEBUG nova.network.neutron [req-bf5edcde-5ebe-4911-ab60-29b4e6d6d02b req-f65c40fb-a89b-4ad2-8363-6c66bfbff7ec service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Refreshing network info cache for port b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1378.252464] env[61985]: DEBUG oslo_vmware.api [-] Task: {'id': task-936902, 'name': CreateVM_Task, 'duration_secs': 0.328226} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1378.252464] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Created VM on the ESX host {{(pid=61985) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1378.259827] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1378.260054] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1378.260449] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1378.260733] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69e79d01-39bb-4aff-9ee8-f1fda52cd010 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.264910] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1378.264910] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523c2fc5-8cc0-f876-202e-20055218ee7f" [ 1378.264910] env[61985]: _type = "Task" [ 1378.264910] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1378.271865] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523c2fc5-8cc0-f876-202e-20055218ee7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1378.567996] env[61985]: DEBUG nova.network.neutron [req-bf5edcde-5ebe-4911-ab60-29b4e6d6d02b req-f65c40fb-a89b-4ad2-8363-6c66bfbff7ec service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updated VIF entry in instance network info cache for port b1e4bff2-85a4-4a99-aa99-072688a294b5. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1378.568376] env[61985]: DEBUG nova.network.neutron [req-bf5edcde-5ebe-4911-ab60-29b4e6d6d02b req-f65c40fb-a89b-4ad2-8363-6c66bfbff7ec service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updating instance_info_cache with network_info: [{"id": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "address": "fa:16:3e:67:e8:13", "network": {"id": "e5781e12-b404-4c77-ab10-4eacacdb6121", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-210244505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f73280838cc84d8b91b13a96d51fad53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e4bff2-85", "ovs_interfaceid": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1378.776339] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]523c2fc5-8cc0-f876-202e-20055218ee7f, 'name': SearchDatastore_Task, 'duration_secs': 0.009009} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1378.776593] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1378.776834] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1378.777090] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1378.777242] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1378.777422] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1378.777677] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b96c0ada-930f-4c9e-9256-39ceb0e43955 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.785318] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1378.785497] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1378.786215] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c9ca163-ac62-4b27-b69e-6ab4922ad0cc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1378.790835] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1378.790835] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a8ea19-eb5c-3358-80ba-de29a551e622" [ 1378.790835] env[61985]: _type = "Task" [ 1378.790835] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1378.797664] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a8ea19-eb5c-3358-80ba-de29a551e622, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.070880] env[61985]: DEBUG oslo_concurrency.lockutils [req-bf5edcde-5ebe-4911-ab60-29b4e6d6d02b req-f65c40fb-a89b-4ad2-8363-6c66bfbff7ec service nova] Releasing lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1379.301506] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52a8ea19-eb5c-3358-80ba-de29a551e622, 'name': SearchDatastore_Task, 'duration_secs': 0.008043} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.302265] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c78f7bd-9e11-4c00-804e-0b9799c21b81 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.306929] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1379.306929] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526f2246-349b-6b21-5496-972d33fff088" [ 1379.306929] env[61985]: _type = "Task" [ 1379.306929] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.313940] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526f2246-349b-6b21-5496-972d33fff088, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1379.816972] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]526f2246-349b-6b21-5496-972d33fff088, 'name': SearchDatastore_Task, 'duration_secs': 0.008546} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1379.817260] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1379.817520] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 752346f6-9093-4ed3-9722-94fa0ccdda98/752346f6-9093-4ed3-9722-94fa0ccdda98.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1379.817800] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc10d9da-4371-4822-9c49-d8c0fe418ef9 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1379.824294] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1379.824294] env[61985]: value = "task-936903" [ 1379.824294] env[61985]: _type = "Task" [ 1379.824294] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1379.831365] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936903, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.334620] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936903, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44829} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1380.337030] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 752346f6-9093-4ed3-9722-94fa0ccdda98/752346f6-9093-4ed3-9722-94fa0ccdda98.vmdk {{(pid=61985) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1380.337030] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Extending root virtual disk to 1048576 {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1380.337030] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd7842f6-5cad-408d-a27d-e68cd908d541 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.342607] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1380.342607] env[61985]: value = "task-936904" [ 1380.342607] env[61985]: _type = "Task" [ 1380.342607] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1380.349577] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936904, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1380.853418] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936904, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063513} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1380.853685] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Extended root virtual disk {{(pid=61985) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1380.854462] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c595602a-eceb-48b6-bf40-b1dde0fae5e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.876835] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 752346f6-9093-4ed3-9722-94fa0ccdda98/752346f6-9093-4ed3-9722-94fa0ccdda98.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1380.877137] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ba017ff-ecfc-460a-b45b-327fb9bbd5e1 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.896329] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1380.896329] env[61985]: value = "task-936905" [ 1380.896329] env[61985]: _type = "Task" [ 1380.896329] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1380.903725] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936905, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1381.407056] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936905, 'name': ReconfigVM_Task, 'duration_secs': 0.325252} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1381.407056] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 752346f6-9093-4ed3-9722-94fa0ccdda98/752346f6-9093-4ed3-9722-94fa0ccdda98.vmdk or device None with type sparse {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1381.407056] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c02e178-5b5f-408e-be96-6e2809de3b0f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.412806] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1381.412806] env[61985]: value = "task-936906" [ 1381.412806] env[61985]: _type = "Task" [ 1381.412806] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1381.419713] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936906, 'name': Rename_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1381.487463] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1381.922736] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936906, 'name': Rename_Task, 'duration_secs': 0.136574} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1381.923027] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1381.923290] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03b568a2-ea16-4661-8b83-a6d37c72914f {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.929968] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1381.929968] env[61985]: value = "task-936907" [ 1381.929968] env[61985]: _type = "Task" [ 1381.929968] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1381.938162] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936907, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1382.439378] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936907, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1382.491995] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.492221] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.492385] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.492524] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61985) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 1382.940925] env[61985]: DEBUG oslo_vmware.api [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936907, 'name': PowerOnVM_Task, 'duration_secs': 0.690015} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1382.941212] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1382.941422] env[61985]: INFO nova.compute.manager [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Took 7.26 seconds to spawn the instance on the hypervisor. [ 1382.941606] env[61985]: DEBUG nova.compute.manager [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1382.942370] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12cffc25-845b-4334-b3dc-1d07f0098f7a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.458775] env[61985]: INFO nova.compute.manager [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Took 12.97 seconds to build instance. [ 1383.491401] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1383.491589] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Starting heal instance info cache {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1383.491661] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Rebuilding the list of instances to heal {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9936}} [ 1383.765064] env[61985]: INFO nova.compute.manager [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Rescuing [ 1383.765306] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquiring lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1383.765452] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquired lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1383.765631] env[61985]: DEBUG nova.network.neutron [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Building network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1383.960447] env[61985]: DEBUG oslo_concurrency.lockutils [None req-72568c2f-8adc-4c0a-8372-60bf47db49ce tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lock "752346f6-9093-4ed3-9722-94fa0ccdda98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.483s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1383.994514] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1384.483521] env[61985]: DEBUG nova.network.neutron [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updating instance_info_cache with network_info: [{"id": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "address": "fa:16:3e:67:e8:13", "network": {"id": "e5781e12-b404-4c77-ab10-4eacacdb6121", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-210244505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f73280838cc84d8b91b13a96d51fad53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e4bff2-85", "ovs_interfaceid": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1384.986050] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Releasing lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1384.989028] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquired lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1384.989028] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Forcefully refreshing network info cache for instance {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1384.989028] env[61985]: DEBUG nova.objects.instance [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lazy-loading 'info_cache' on Instance uuid 752346f6-9093-4ed3-9722-94fa0ccdda98 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1385.515718] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1385.516231] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc955b2a-17ab-4bde-a15d-51e0faac7054 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.523787] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1385.523787] env[61985]: value = "task-936908" [ 1385.523787] env[61985]: _type = "Task" [ 1385.523787] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1385.532295] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936908, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.034383] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936908, 'name': PowerOffVM_Task, 'duration_secs': 0.196697} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1386.035191] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1386.035960] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c965352-69b8-4127-ae11-2b0c1b82f123 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.056301] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7fcaa9-8bf2-4332-9340-9ba8650ee20d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.081331] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1386.081777] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25b0f678-29aa-4230-a9dd-0cba37669bf3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.088740] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1386.088740] env[61985]: value = "task-936909" [ 1386.088740] env[61985]: _type = "Task" [ 1386.088740] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1386.095949] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936909, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.600472] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] VM already powered off {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1386.600793] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Processing image fe6d817b-6194-440f-988a-f4a94c580922 {{(pid=61985) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1386.600947] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1386.601126] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1386.601313] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1386.601562] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c7f54138-905a-43a3-9af5-39dc09e25fba {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.609577] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61985) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1386.609759] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61985) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1386.610484] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e1c99f3-88de-433b-a960-3394a76cca77 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.615266] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1386.615266] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f15186-754c-3c20-1ebf-52696329c1b0" [ 1386.615266] env[61985]: _type = "Task" [ 1386.615266] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1386.622204] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f15186-754c-3c20-1ebf-52696329c1b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.735323] env[61985]: DEBUG nova.network.neutron [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updating instance_info_cache with network_info: [{"id": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "address": "fa:16:3e:67:e8:13", "network": {"id": "e5781e12-b404-4c77-ab10-4eacacdb6121", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-210244505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f73280838cc84d8b91b13a96d51fad53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e4bff2-85", "ovs_interfaceid": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1387.126037] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52f15186-754c-3c20-1ebf-52696329c1b0, 'name': SearchDatastore_Task, 'duration_secs': 0.01127} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1387.126834] env[61985]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a12922bd-e2d5-4bcd-a7f9-483549e9e1dc {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.132174] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1387.132174] env[61985]: value = "session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52433d68-ae16-6383-6a7f-1daf6609bfab" [ 1387.132174] env[61985]: _type = "Task" [ 1387.132174] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1387.139864] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52433d68-ae16-6383-6a7f-1daf6609bfab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1387.238623] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Releasing lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1387.238923] env[61985]: DEBUG nova.compute.manager [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updated the network info_cache for instance {{(pid=61985) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10003}} [ 1387.239200] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.239412] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.239601] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.642336] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': session[524dc94e-1cc7-81b3-1a90-53e590c07f40]52433d68-ae16-6383-6a7f-1daf6609bfab, 'name': SearchDatastore_Task, 'duration_secs': 0.009962} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1387.642724] env[61985]: DEBUG oslo_concurrency.lockutils [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1387.642930] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 752346f6-9093-4ed3-9722-94fa0ccdda98/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk. {{(pid=61985) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1387.643247] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c8fe544-33a8-4bb6-be32-14824f70f5d0 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.649879] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1387.649879] env[61985]: value = "task-936910" [ 1387.649879] env[61985]: _type = "Task" [ 1387.649879] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1387.656943] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936910, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1387.743335] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1387.743587] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1387.743733] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1387.743891] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61985) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1387.744794] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6047a0da-4734-4947-8daa-d4812e3d415e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.753852] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66651d2-f280-480e-9311-625dfd2bcac4 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.767309] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d5aa9b-e871-495a-b740-3d82fc00ea42 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.773412] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22625d3-0790-4e1f-a660-966e5fd50110 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.802233] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181386MB free_disk=43GB free_vcpus=48 pci_devices=None {{(pid=61985) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1387.802445] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1387.802594] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1388.160644] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936910, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458584} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1388.160906] env[61985]: INFO nova.virt.vmwareapi.ds_util [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fe6d817b-6194-440f-988a-f4a94c580922/fe6d817b-6194-440f-988a-f4a94c580922.vmdk to [datastore2] 752346f6-9093-4ed3-9722-94fa0ccdda98/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk. [ 1388.161733] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060cef92-56ac-4902-8504-260f9d0e2013 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.186996] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 752346f6-9093-4ed3-9722-94fa0ccdda98/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1388.187306] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cb4cf73-fe5d-4548-8d26-270a3e41311d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.205559] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1388.205559] env[61985]: value = "task-936911" [ 1388.205559] env[61985]: _type = "Task" [ 1388.205559] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1388.213668] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936911, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1388.715050] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936911, 'name': ReconfigVM_Task, 'duration_secs': 0.296903} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1388.715448] env[61985]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 752346f6-9093-4ed3-9722-94fa0ccdda98/fe6d817b-6194-440f-988a-f4a94c580922-rescue.vmdk or device None with type thin {{(pid=61985) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1388.716269] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c8ad43-fd6e-417d-b208-ebef7a0c8820 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.740716] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-547dfb34-3aed-471b-bf81-cbf1b0950274 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.755491] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1388.755491] env[61985]: value = "task-936912" [ 1388.755491] env[61985]: _type = "Task" [ 1388.755491] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1388.762756] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936912, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1388.826182] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Instance 752346f6-9093-4ed3-9722-94fa0ccdda98 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61985) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1388.826476] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1388.826702] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61985) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1388.853951] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236e1aae-fb90-48e0-9a1b-a867ceae2787 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.861569] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d341ab09-d770-4299-b663-b4bfa1f1264a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.891165] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bfebeba-8bf5-4920-988b-b30729236baf {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.898178] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c8a8a3-e09e-4208-8b30-1c06cfef1fd3 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.911480] env[61985]: DEBUG nova.compute.provider_tree [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1389.266714] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936912, 'name': ReconfigVM_Task, 'duration_secs': 0.148405} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1389.266950] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Powering on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1389.267218] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2287f82f-fab4-48b3-8328-ff0fdc3caf5a {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.273813] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1389.273813] env[61985]: value = "task-936913" [ 1389.273813] env[61985]: _type = "Task" [ 1389.273813] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1389.280742] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936913, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1389.414151] env[61985]: DEBUG nova.scheduler.client.report [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1389.784211] env[61985]: DEBUG oslo_vmware.api [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936913, 'name': PowerOnVM_Task, 'duration_secs': 0.367779} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1389.784635] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Powered on the VM {{(pid=61985) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1389.787551] env[61985]: DEBUG nova.compute.manager [None req-5a200884-2d5f-4d00-9c37-fbfb4dae7664 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Checking state {{(pid=61985) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1389.788314] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a2465a1-a67c-4c46-b13f-f8f5804ef45d {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.919166] env[61985]: DEBUG nova.compute.resource_tracker [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61985) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1389.919416] env[61985]: DEBUG oslo_concurrency.lockutils [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.117s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1390.541762] env[61985]: DEBUG nova.compute.manager [req-923f2a83-ac60-4359-a4df-682a2a75a1dd req-aec2e73c-4930-4dd8-a101-d236337ad20a service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Received event network-changed-b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1390.542043] env[61985]: DEBUG nova.compute.manager [req-923f2a83-ac60-4359-a4df-682a2a75a1dd req-aec2e73c-4930-4dd8-a101-d236337ad20a service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Refreshing instance network info cache due to event network-changed-b1e4bff2-85a4-4a99-aa99-072688a294b5. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1390.542332] env[61985]: DEBUG oslo_concurrency.lockutils [req-923f2a83-ac60-4359-a4df-682a2a75a1dd req-aec2e73c-4930-4dd8-a101-d236337ad20a service nova] Acquiring lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.542520] env[61985]: DEBUG oslo_concurrency.lockutils [req-923f2a83-ac60-4359-a4df-682a2a75a1dd req-aec2e73c-4930-4dd8-a101-d236337ad20a service nova] Acquired lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.542737] env[61985]: DEBUG nova.network.neutron [req-923f2a83-ac60-4359-a4df-682a2a75a1dd req-aec2e73c-4930-4dd8-a101-d236337ad20a service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Refreshing network info cache for port b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1391.171348] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1391.317932] env[61985]: DEBUG nova.network.neutron [req-923f2a83-ac60-4359-a4df-682a2a75a1dd req-aec2e73c-4930-4dd8-a101-d236337ad20a service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updated VIF entry in instance network info cache for port b1e4bff2-85a4-4a99-aa99-072688a294b5. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1391.318380] env[61985]: DEBUG nova.network.neutron [req-923f2a83-ac60-4359-a4df-682a2a75a1dd req-aec2e73c-4930-4dd8-a101-d236337ad20a service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updating instance_info_cache with network_info: [{"id": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "address": "fa:16:3e:67:e8:13", "network": {"id": "e5781e12-b404-4c77-ab10-4eacacdb6121", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-210244505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f73280838cc84d8b91b13a96d51fad53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e4bff2-85", "ovs_interfaceid": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1391.821708] env[61985]: DEBUG oslo_concurrency.lockutils [req-923f2a83-ac60-4359-a4df-682a2a75a1dd req-aec2e73c-4930-4dd8-a101-d236337ad20a service nova] Releasing lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1392.564929] env[61985]: DEBUG nova.compute.manager [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Received event network-changed-b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1392.565179] env[61985]: DEBUG nova.compute.manager [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Refreshing instance network info cache due to event network-changed-b1e4bff2-85a4-4a99-aa99-072688a294b5. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1392.565381] env[61985]: DEBUG oslo_concurrency.lockutils [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] Acquiring lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1392.565530] env[61985]: DEBUG oslo_concurrency.lockutils [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] Acquired lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1392.565696] env[61985]: DEBUG nova.network.neutron [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Refreshing network info cache for port b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1393.374467] env[61985]: DEBUG nova.network.neutron [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updated VIF entry in instance network info cache for port b1e4bff2-85a4-4a99-aa99-072688a294b5. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1393.374827] env[61985]: DEBUG nova.network.neutron [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updating instance_info_cache with network_info: [{"id": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "address": "fa:16:3e:67:e8:13", "network": {"id": "e5781e12-b404-4c77-ab10-4eacacdb6121", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-210244505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f73280838cc84d8b91b13a96d51fad53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e4bff2-85", "ovs_interfaceid": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1393.878009] env[61985]: DEBUG oslo_concurrency.lockutils [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] Releasing lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1393.878009] env[61985]: DEBUG nova.compute.manager [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Received event network-changed-b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1393.878009] env[61985]: DEBUG nova.compute.manager [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Refreshing instance network info cache due to event network-changed-b1e4bff2-85a4-4a99-aa99-072688a294b5. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1393.878380] env[61985]: DEBUG oslo_concurrency.lockutils [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] Acquiring lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1393.878380] env[61985]: DEBUG oslo_concurrency.lockutils [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] Acquired lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1393.878531] env[61985]: DEBUG nova.network.neutron [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Refreshing network info cache for port b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1394.516448] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquiring lock "752346f6-9093-4ed3-9722-94fa0ccdda98" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1394.516885] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lock "752346f6-9093-4ed3-9722-94fa0ccdda98" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.517263] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquiring lock "752346f6-9093-4ed3-9722-94fa0ccdda98-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1394.517600] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lock "752346f6-9093-4ed3-9722-94fa0ccdda98-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.517935] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lock "752346f6-9093-4ed3-9722-94fa0ccdda98-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1394.520979] env[61985]: INFO nova.compute.manager [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Terminating instance [ 1394.523506] env[61985]: DEBUG nova.compute.manager [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Start destroying the instance on the hypervisor. {{(pid=61985) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3156}} [ 1394.523840] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Destroying instance {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1394.525169] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ee8420-0094-4d79-8db7-ca92664548f2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.539081] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Powering off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1394.539438] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-646be5de-3afa-4833-a888-d9b0b105f326 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.546743] env[61985]: DEBUG oslo_vmware.api [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1394.546743] env[61985]: value = "task-936914" [ 1394.546743] env[61985]: _type = "Task" [ 1394.546743] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1394.557965] env[61985]: DEBUG oslo_vmware.api [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936914, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1394.564918] env[61985]: DEBUG nova.network.neutron [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updated VIF entry in instance network info cache for port b1e4bff2-85a4-4a99-aa99-072688a294b5. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1394.565310] env[61985]: DEBUG nova.network.neutron [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updating instance_info_cache with network_info: [{"id": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "address": "fa:16:3e:67:e8:13", "network": {"id": "e5781e12-b404-4c77-ab10-4eacacdb6121", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-210244505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f73280838cc84d8b91b13a96d51fad53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e4bff2-85", "ovs_interfaceid": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1394.589591] env[61985]: DEBUG nova.compute.manager [req-582a7958-566f-40fb-8681-c646722e4754 req-53e8e3aa-9747-426b-957b-5580be3a7cfd service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Received event network-changed-b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1394.589847] env[61985]: DEBUG nova.compute.manager [req-582a7958-566f-40fb-8681-c646722e4754 req-53e8e3aa-9747-426b-957b-5580be3a7cfd service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Refreshing instance network info cache due to event network-changed-b1e4bff2-85a4-4a99-aa99-072688a294b5. {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11140}} [ 1394.589961] env[61985]: DEBUG oslo_concurrency.lockutils [req-582a7958-566f-40fb-8681-c646722e4754 req-53e8e3aa-9747-426b-957b-5580be3a7cfd service nova] Acquiring lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1395.056095] env[61985]: DEBUG oslo_vmware.api [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936914, 'name': PowerOffVM_Task, 'duration_secs': 0.197027} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1395.056442] env[61985]: DEBUG nova.virt.vmwareapi.vm_util [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Powered off the VM {{(pid=61985) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1395.056583] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Unregistering the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1395.056867] env[61985]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e492e6e-8f5a-42e7-96da-381725ecc0a6 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.068068] env[61985]: DEBUG oslo_concurrency.lockutils [req-894c8efd-8f22-4a70-ac57-b67dfc8ee6be req-f003f2f7-dfff-4d21-8aa5-71984b3db6a7 service nova] Releasing lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1395.068512] env[61985]: DEBUG oslo_concurrency.lockutils [req-582a7958-566f-40fb-8681-c646722e4754 req-53e8e3aa-9747-426b-957b-5580be3a7cfd service nova] Acquired lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1395.068752] env[61985]: DEBUG nova.network.neutron [req-582a7958-566f-40fb-8681-c646722e4754 req-53e8e3aa-9747-426b-957b-5580be3a7cfd service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Refreshing network info cache for port b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1395.118716] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Unregistered the VM {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1395.118931] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Deleting contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1395.119135] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Deleting the datastore file [datastore2] 752346f6-9093-4ed3-9722-94fa0ccdda98 {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1395.119399] env[61985]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1fc3340-33d1-42a7-8efd-80adde5361ac {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.125771] env[61985]: DEBUG oslo_vmware.api [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for the task: (returnval){ [ 1395.125771] env[61985]: value = "task-936916" [ 1395.125771] env[61985]: _type = "Task" [ 1395.125771] env[61985]: } to complete. {{(pid=61985) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1395.133477] env[61985]: DEBUG oslo_vmware.api [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936916, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1395.638766] env[61985]: DEBUG oslo_vmware.api [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Task: {'id': task-936916, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167189} completed successfully. {{(pid=61985) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1395.638766] env[61985]: DEBUG nova.virt.vmwareapi.ds_util [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Deleted the datastore file {{(pid=61985) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1395.638766] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Deleted contents of the VM from datastore datastore2 {{(pid=61985) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1395.638766] env[61985]: DEBUG nova.virt.vmwareapi.vmops [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Instance destroyed {{(pid=61985) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1395.638766] env[61985]: INFO nova.compute.manager [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1395.638766] env[61985]: DEBUG oslo.service.loopingcall [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61985) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1395.638766] env[61985]: DEBUG nova.compute.manager [-] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Deallocating network for instance {{(pid=61985) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1395.638766] env[61985]: DEBUG nova.network.neutron [-] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] deallocate_for_instance() {{(pid=61985) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1395.738017] env[61985]: DEBUG nova.network.neutron [req-582a7958-566f-40fb-8681-c646722e4754 req-53e8e3aa-9747-426b-957b-5580be3a7cfd service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updated VIF entry in instance network info cache for port b1e4bff2-85a4-4a99-aa99-072688a294b5. {{(pid=61985) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1395.738399] env[61985]: DEBUG nova.network.neutron [req-582a7958-566f-40fb-8681-c646722e4754 req-53e8e3aa-9747-426b-957b-5580be3a7cfd service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updating instance_info_cache with network_info: [{"id": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "address": "fa:16:3e:67:e8:13", "network": {"id": "e5781e12-b404-4c77-ab10-4eacacdb6121", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-210244505-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f73280838cc84d8b91b13a96d51fad53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1e4bff2-85", "ovs_interfaceid": "b1e4bff2-85a4-4a99-aa99-072688a294b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1395.875039] env[61985]: DEBUG nova.compute.manager [req-656ff58d-5a5d-4353-bf3b-a58394652afc req-f190771e-5ef1-436e-80cd-d2c1d39423d3 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Received event network-vif-deleted-b1e4bff2-85a4-4a99-aa99-072688a294b5 {{(pid=61985) external_instance_event /opt/stack/nova/nova/compute/manager.py:11135}} [ 1395.875267] env[61985]: INFO nova.compute.manager [req-656ff58d-5a5d-4353-bf3b-a58394652afc req-f190771e-5ef1-436e-80cd-d2c1d39423d3 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Neutron deleted interface b1e4bff2-85a4-4a99-aa99-072688a294b5; detaching it from the instance and deleting it from the info cache [ 1395.875449] env[61985]: DEBUG nova.network.neutron [req-656ff58d-5a5d-4353-bf3b-a58394652afc req-f190771e-5ef1-436e-80cd-d2c1d39423d3 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1396.240931] env[61985]: DEBUG oslo_concurrency.lockutils [req-582a7958-566f-40fb-8681-c646722e4754 req-53e8e3aa-9747-426b-957b-5580be3a7cfd service nova] Releasing lock "refresh_cache-752346f6-9093-4ed3-9722-94fa0ccdda98" {{(pid=61985) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1396.354719] env[61985]: DEBUG nova.network.neutron [-] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Updating instance_info_cache with network_info: [] {{(pid=61985) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1396.378139] env[61985]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8477fb5c-8bd9-489f-a941-b6dcc8e0e901 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.387775] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae1d81e-b8f9-4950-a232-37cf6e0e3d3e {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.409818] env[61985]: DEBUG nova.compute.manager [req-656ff58d-5a5d-4353-bf3b-a58394652afc req-f190771e-5ef1-436e-80cd-d2c1d39423d3 service nova] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Detach interface failed, port_id=b1e4bff2-85a4-4a99-aa99-072688a294b5, reason: Instance 752346f6-9093-4ed3-9722-94fa0ccdda98 could not be found. {{(pid=61985) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10969}} [ 1396.857228] env[61985]: INFO nova.compute.manager [-] [instance: 752346f6-9093-4ed3-9722-94fa0ccdda98] Took 1.22 seconds to deallocate network for instance. [ 1397.364639] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1397.364957] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1397.365212] env[61985]: DEBUG nova.objects.instance [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lazy-loading 'resources' on Instance uuid 752346f6-9093-4ed3-9722-94fa0ccdda98 {{(pid=61985) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1397.904123] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d352ac05-1779-4af7-913d-4ea4e10511d2 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.911917] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b102a9-6b55-4d2e-a9af-509e2b08bc5b {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.941425] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a469a7d-157d-45b5-b7f9-729204341676 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.947974] env[61985]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a85dbb-6ff5-4fb9-8f44-00164ddfe8e7 {{(pid=61985) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.961631] env[61985]: DEBUG nova.compute.provider_tree [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Inventory has not changed in ProviderTree for provider: aed7e5b3-c662-4538-8447-c4f67b460215 {{(pid=61985) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1398.464240] env[61985]: DEBUG nova.scheduler.client.report [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Inventory has not changed for provider aed7e5b3-c662-4538-8447-c4f67b460215 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 43, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61985) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1398.486664] env[61985]: DEBUG oslo_service.periodic_task [None req-43467a38-4e74-4b85-b948-27fd9ad666d6 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61985) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1398.969896] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.605s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1398.988315] env[61985]: INFO nova.scheduler.client.report [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Deleted allocations for instance 752346f6-9093-4ed3-9722-94fa0ccdda98 [ 1399.496486] env[61985]: DEBUG oslo_concurrency.lockutils [None req-c51c9138-3a0e-4a21-b485-38364e8a4475 tempest-ServerRescueTestJSONUnderV235-2089276474 tempest-ServerRescueTestJSONUnderV235-2089276474-project-member] Lock "752346f6-9093-4ed3-9722-94fa0ccdda98" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.980s {{(pid=61985) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}